-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathlibrary.bib
1720 lines (1715 loc) · 214 KB
/
library.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
Automatically generated by Mendeley Desktop 1.19.8
Any changes to this file will be lost if it is regenerated by Mendeley.
BibTeX export options can be customized via Options -> BibTeX in Mendeley Desktop
@book{IanGoodfellow2016,
abstract = {"Deep learning is a form of machine learning that enables computers to learn from experience and understand the world in terms of a hierarchy of concepts. Because the computer gathers knowledge from experience, there is no need for a human computer operator to formally specify all the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones; a graph of these hierarchies would be many layers deep. This book introduces a broad range of topics in deep learning. The text offers mathematical and conceptual background, covering relevant concepts in linear algebra, probability theory and information theory, numerical computation, and machine learning. It describes deep learning techniques used by practitioners in industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling, and practical methodology; and it surveys such applications as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and video games. Finally, the book offers research perspectives, covering such theoretical topics as linear factor models, autoencoders, representation learning, structured probabilistic models, Monte Carlo methods, the partition function, approximate inference, and deep generative models. Deep Learning can be used by undergraduate or graduate students planning careers in either industry or research, and by software engineers who want to begin using deep learning in their products or platforms. A website offers supplementary material for both readers and instructors"--Page 4 of cover. Introduction -- Applied math and machine learning basics. Linear algebra -- Probability and information theory -- Numerical computation -- Machine learning basics -- Deep networks: modern practices. Deep feedforward networks -- Regularization for deep learning -- Optimization for training deep models -- Convolutional networks -- Sequence modeling: recurrent and recursive nets -- Practical methodology -- Applications -- Deep learning research. Linear factor models -- Autoencoders -- Representation learning -- Structured probabilistic models for deep learning -- Monte Carlo methods -- Confronting the partition function -- Approximate inference -- Deep generative models.},
author = {Courville, Ian Goodfellow and Yoshua Bengio and Aaron},
booktitle = {Nature},
isbn = {3463353563306},
keywords = {cnn,convolutional neural networks,deep learning,plant disease},
number = {7553},
pages = {1--73},
publisher = {MIT Press},
title = {{Deep learning}},
url = {https://www.deeplearningbook.org/},
volume = {29},
year = {2016}
}
@article{Kristjanpoller2018,
abstract = {Measurement, prediction, and modeling of currency price volatility constitutes an important area of research at both the national and corporate level. Countries attempt to understand currency volatility to set national economic policies and firms to best manage exchange rate risk and leverage assets. A relatively new technological invention that the corporate treasurer has to turn to as part of the overall financial strategy is cryptocurrency. One estimate values the total market capitalization of cryptocurrencies at {\$}557 billion USD at the beginning of 2018. While the overall size of the market for cryptocurrency is significant, our understanding of the behavior of this instrument is only beginning. In this article, we propose a hybrid Artificial Neural Network-Generalized AutoRegressive Conditional Heteroskedasticity (ANN-GARCH) model with preprocessing to forecast the price volatility of bitcoin, the most traded and largest by market capitalization of the cryptocurrencies.},
author = {Kristjanpoller, Werner and Minutolo, Marcel C.},
doi = {10.1016/j.eswa.2018.05.011},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Kristjanpoller, Minutolo - 2018 - A hybrid volatility forecasting framework integrating GARCH, artificial neural network, technical anal.pdf:pdf},
issn = {09574174},
journal = {Expert Syst. Appl.},
title = {{A hybrid volatility forecasting framework integrating GARCH, artificial neural network, technical analysis and principal components analysis}},
year = {2018}
}
@article{Russel2015,
abstract = {A rtificial intelligence (AI) research has explored a variety of problems and approaches since its inception, but for the last 20 years or so has been focused on the problems surrounding the construction of intelligent agents-systems that perceive and act in some environment. In this context, the criterion for intelligence is related to statistical and economic notions of rationality-colloquially, the ability to make good decisions, plans, or inferences. The adoption of probabilistic representations and statistical learning methods has led to a large degree of integration and cross-fertilization between AI, machine learning, statistics, control theory, neuroscience, and other fields. The establishment of shared theoretical frameworks, combined with the availability of data and processing power, has yielded remarkable successes in various component tasks such as speech recognition , image classification, autonomous vehicles, machine translation, legged locomotion, and question-answering systems .},
author = {Russell, Stuart and Dewey, Daniel and Tegmark, Max},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Russell, Dewey, Tegmark - 2015 - Research Priorities for Robust and Beneficial Artificial Intelligence(2).pdf:pdf},
issn = {0738-4602},
title = {{Research Priorities for Robust and Beneficial Artificial Intelligence}},
year = {2015}
}
@techreport{Cheriton,
abstract = {The amount of data generated in the modern society is increasing rapidly. New problems and novel approaches of data capture, storage, analysis and visualization are responsible for the emergence of the Big Data research field. Machine Learning algorithms can be used in Big Data to make better and more accurate inferences. However, because of the challenges Big Data imposes, these algorithms need to be adapted and optimized to specific applications. One important decision made by software engineers is the choice of the language that is used in the implementation of these algorithms. Therefore, this literature survey identifies and describes domain-specific languages and frameworks used for Machine Learning in Big Data. By doing this, software engineers can then make more informed choices and beginners have an overview of the main languages used in this domain.},
author = {Cheriton, David R and Alencar, Paulo and Cowan, Donald},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Cheriton, Alencar, Cowan - Unknown - A Survey on Domain-Specific Languages for Machine Learning in Big Data Ivens Portugal.pdf:pdf},
keywords = {BD,Big Data,DSL,ML,Machine Learning,domain-specific languages,literature survey},
title = {{A Survey on Domain-Specific Languages for Machine Learning in Big Data Ivens Portugal}}
}
@techreport{Chen,
abstract = {As machine learning (ML) has seen increasing adoption in safety-critical domains (e.g., autonomous vehicles), the reliability of ML systems has also grown in importance. While prior studies have proposed techniques to enable efficient error-resilience techniques (e.g., selective instruction duplication), a fundamental requirement for realizing these techniques is a detailed understanding of the application's resilience. In this work, we present TensorFI, a high-level fault injection (FI) framework for TensorFlow-based applications. TensorFI is able to inject both hardware and software faults in general TensorFlow programs. TensorFI is a configurable FI tool that is flexible, easy to use, and portable. It can be integrated into existing TensorFlow programs to assess their resilience for different fault types (e.g., faults in particular operators). We use TensorFI to evaluate the resilience of 12 ML programs, including DNNs used in the autonomous vehicle domain. Our tool is publicly available at https://github.com/DependableSystemsLab/TensorFI .},
archivePrefix = {arXiv},
arxivId = {2004.01743v1},
author = {Chen, Zitao and Narayanan, Niranjhana and Fang, Bo and Li, Guanpeng and Pattabiraman, Karthik and Debardeleben, Nathan},
eprint = {2004.01743v1},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Chen et al. - Unknown - TensorFI A Flexible Fault Injection Framework for TensorFlow Applications.pdf:pdf},
title = {{TensorFI: A Flexible Fault Injection Framework for TensorFlow Applications}},
url = {https://github.com/DependableSystemsLab/TensorFI}
}
@techreport{Fisac,
abstract = {Safety analysis is a necessary component in the design and deployment of autonomous systems. Techniques from robust optimal control theory, such as Hamilton-Jacobi reachability analysis, allow a rigorous formalization of safety as guaranteed constraint satisfaction. Unfortunately, the computational complexity of these tools for general dynamical systems scales poorly with state dimension, making existing tools impractical beyond small problems. Modern reinforcement learning methods have shown promising ability to find approximate yet proficient solutions to optimal control problems in complex and high-dimensional systems, however their formulation is restricted to problems with an additive payoff (reward) over time, unsuitable for reasoning about safety. In recent work, we proved that the problem of maximizing the minimum payoff over time, central to safety analysis, can be time-discounted to induce a contraction mapping. Here, we introduce a novel, time-discounted Safety Bellman Equation that renders reinforcement learning techniques amenable to quantitative safety analysis, enabling them to approximate the safe set and optimal safety policy. This opens a new avenue of research connecting control-theoretic safety analysis and the reinforcement learning domain. We demonstrate our formulation on a variety of simulated robotics tasks and reinforcement learning schemes, validating our results against analytic and numerical solutions when these can be obtained, and showing scalability to previously intractable problems of up to 18 state dimensions by exploiting state-of-the-art deep reinforcement learning algorithms.},
author = {Fisac, Jaime F and Lugovoy, Neil F and Rubies-Royo, Vicen{\c{c}} and Ghosh, Shromona and Tomlin, Claire J},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Fisac et al. - Unknown - Bridging Hamilton-Jacobi Safety Analysis and Reinforcement Learning.pdf:pdf},
title = {{Bridging Hamilton-Jacobi Safety Analysis and Reinforcement Learning}}
}
@inproceedings{Bloomfield2010,
abstract = {This paper focuses on the approaches used in safety cases for software based systems. We outline the history of approaches for assuring the safety of software-based systems, the current uptake of safety and assurance cases and the current practice on structured safety cases. Directions for further development are discussed. {\textcopyright} Springer-Verlag London Limited 2010.},
author = {Bloomfield, Robin and Bishop, Peter},
booktitle = {Mak. Syst. Safer - Proc. 18th Safety-Critical Syst. Symp. SSS 2010},
doi = {10.1007/978-1-84996-086-1_4},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Bloomfield, Bishop - 2010 - Safety and assurance cases Past, present and possible future - An adelard perspective.pdf:pdf},
isbn = {9781849960854},
pages = {51--67},
publisher = {Springer London},
title = {{Safety and assurance cases: Past, present and possible future - An adelard perspective}},
year = {2010}
}
@article{Arlot2010,
abstract = {Used to estimate the risk of an estimator or to perform model selection, cross-validation is a widespread strategy because of its simplic-ity and its (apparent) universality. Many results exist on model selection performances of cross-validation procedures. This survey intends to relate these results to the most recent advances of model selection theory, with a particular emphasis on distinguishing empirical statements from rigorous theoretical results. As a conclusion, guidelines are provided for choosing the best cross-validation procedure according to the particular features of the problem in hand.},
archivePrefix = {arXiv},
arxivId = {0907.4728},
author = {Arlot, Sylvain and Celisse, Alain},
doi = {10.1214/09-SS054},
eprint = {0907.4728},
issn = {19357516},
journal = {Stat. Surv.},
keywords = {Cross-validation,Leave-one-out,Model selection},
pages = {40--79},
title = {{A survey of cross-validation procedures for model selection}},
volume = {4},
year = {2010}
}
@inproceedings{Schwalbe2020,
abstract = {Methods for safety assurance suggested by the ISO 26262 automotive functional safety standard are not sufficient for applications based on machine learning (ML). We provide a structured, certification oriented overview on available methods supporting the safety argumen-tation of a ML based system. It is sorted into life-cycle phases, and maturity of the approach as well as applicability to different ML types are collected. From this we deduce current open challenges: powerful solvers, inclusion of expert knowledge, validation of data representativity and model diversity, and model introspection with provable guarantees.},
author = {Schwalbe, Gesina and Schels, Martin},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Schwalbe, Schels - 2020 - A Survey on Methods for the Safety Assurance of Machine Learning Based Systems A Survey on Methods for the (2).pdf:pdf},
keywords = {ISO 26262,explainable AI,functional safety,life-cycle,machine learning},
title = {{A Survey on Methods for the Safety Assurance of Machine Learning Based Systems}},
url = {https://hal.archives-ouvertes.fr/hal-02442819},
year = {2020}
}
@article{iso15026-1-2019,
doi = {10.1109/IEEESTD.2019.8657410},
file = {:C$\backslash$:/Users/Suchendra/OneDrive - McMaster University/PhD Thesis/Books/15026-1-2019.pdf:pdf},
journal = {ISO/IEC/IEEE 15026-1:2019(E)},
pages = {1--38},
title = {{15026-1-2019 - ISO/IEC/IEEE International Standard - Systems and software engineering--Systems and software assurance --Part 1:Concepts and vocabulary}},
year = {2019}
}
@techreport{Ries,
author = {Ries, Benoˆıt Benoˆıt and Guelfi, Nicolas and Jahi´c, Benjamin Jahi´c},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Ries, Guelfi, Jahi´c - Unknown - An MDE Method for Improving Deep Learning Dataset Requirements Engineering using Alloy and UML.pdf:pdf},
keywords = {Alloy,EMF,Model-Driven Engineering,Requirements Engineering,Sirius,Software Engineering},
title = {{An MDE Method for Improving Deep Learning Dataset Requirements Engineering using Alloy and UML}},
url = {https://orcid.org/0000-0002-8680-2797}
}
@article{Lage2018,
abstract = {We often desire our models to be interpretable as well as accurate. Prior work on optimizing models for interpretability has relied on easy-to-quantify proxies for interpretability, such as sparsity or the number of operations required. In this work, we optimize for interpretability by directly including humans in the optimization loop. We develop an algorithm that minimizes the number of user studies to find models that are both predictive and interpretable and demonstrate our approach on several data sets. Our human subjects results show trends towards different proxy notions of interpretability on different datasets, which suggests that different proxies are preferred on different tasks.},
archivePrefix = {arXiv},
arxivId = {1805.11571},
author = {Lage, Isaac and Ross, Andrew Slavin and Kim, Been and Gershman, Samuel J. and Doshi-Velez, Finale},
eprint = {1805.11571},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Lage et al. - 2018 - Human-in-the-Loop Interpretability Prior.pdf:pdf},
journal = {Adv. Neural Inf. Process. Syst.},
month = {may},
pages = {10159--10168},
publisher = {Neural information processing systems foundation},
title = {{Human-in-the-Loop Interpretability Prior}},
url = {http://arxiv.org/abs/1805.11571},
volume = {2018-December},
year = {2018}
}
@inproceedings{DAlterio2020,
abstract = {In recent year, there has been a growing need for intelligent systems that not only are able to provide reliable classifications but can also produce explanations for the decisions they make. The demand for increased explainability has led to the emergence of explainable artificial intelligence (XAI) as a specific research field. In this context, fuzzy logic systems represent a promising tool thanks to their inherently interpretable structure. The use of a rule-base and linguistic terms, in fact, have allowed researchers to create models that are able to produce explanations in natural language for each of the classifications they make. So far, however, designing systems that make use of interval type-2 (IT2) fuzzy logic and also give explanations for their outputs has been very challenging, partially due to the presence of the type-reduction step. In this paper, it will be shown how constrained interval type-2 (CIT2) fuzzy sets represent a valid alternative to conventional interval type-2 sets in order to address this issue. Through the analysis of two case studies from the medical domain, it is shown how explainable CIT2 classifiers are produced. These systems can explain which rules contributed to the creation of each of the endpoints of the output interval centroid, while showing (in these examples) the same level of accuracy as their IT2 counterpart.},
author = {D'Alterio, Pasquale and Garibaldi, Jonathan M. and John, Robert I.},
booktitle = {IEEE Int. Conf. Fuzzy Syst.},
doi = {10.1109/FUZZ48607.2020.9177671},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Constrained Interval Type-2 Fuzzy classification systems for XAI.pdf:pdf},
isbn = {9781728169323},
issn = {10987584},
keywords = {Constrained interval type-2,Explainable type-2 fuzzy systems,XAI},
month = {jul},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Constrained interval type-2 fuzzy classification systems for explainable AI (XAI)}},
volume = {2020-July},
year = {2020}
}
@techreport{Agrawal2019,
abstract = {TensorFlow Eager is a multi-stage, Python-embedded domain-specific language for hardware-accelerated machine learning, suitable for both interactive research and production. TensorFlow, which TensorFlow Eager extends, requires users to represent computations as dataflow graphs; this permits compiler optimizations and simplifies deployment but hinders rapid prototyping and run-time dynamism. TensorFlow Eager eliminates these usability costs without sacrificing the benefits furnished by graphs: It provides an imperative front-end to TensorFlow that executes operations immediately and a JIT tracer that translates Python functions composed of TensorFlow operations into executable dataflow graphs. TensorFlow Eager thus offers a multi-stage programming model that makes it easy to interpolate between imperative and staged execution in a single package.},
archivePrefix = {arXiv},
arxivId = {1903.01855v1},
author = {Agrawal, Akshay and Modi, Akshay Naresh and Passos, Alexandre and Lavoie, Allen and Agarwal, Ashish and Shankar, Asim and Ganichev, Igor and Levenberg, Josh and Hong, Mingsheng and Monga, Rajat and Cai, Shanqing},
eprint = {1903.01855v1},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Agrawal et al. - 2019 - TENSORFLOW EAGER A MULTI-STAGE, PYTHON-EMBEDDED DSL FOR MACHINE LEARNING.pdf:pdf},
title = {{TENSORFLOW EAGER: A MULTI-STAGE, PYTHON-EMBEDDED DSL FOR MACHINE LEARNING}},
year = {2019}
}
@techreport{Qayyum,
abstract = {Recent years have witnessed widespread adoption of machine learning (ML)/deep learning (DL) techniques due to their superior performance for a variety of healthcare applications ranging from the prediction of cardiac arrest from one-dimensional heart signals to computer-aided diagnosis (CADx) using multi-dimensional medical images. Notwithstanding the impressive performance of ML/DL, there are still lingering doubts regarding the robustness of ML/DL in healthcare settings (which is traditionally considered quite challenging due to the myriad security and privacy issues involved), especially in light of recent results that have shown that ML/DL are vulnerable to adversarial attacks. In this paper, we present an overview of various application areas in healthcare that leverage such techniques from security and privacy point of view and present associated challenges. In addition, we present potential methods to ensure secure and privacy-preserving ML for healthcare applications. Finally, we provide insight into the current research challenges and promising directions for future research.},
archivePrefix = {arXiv},
arxivId = {2001.08103v1},
author = {Qayyum, Adnan and Qadir, Junaid and Bilal, Muhammad and Al-Fuqaha, Ala},
eprint = {2001.08103v1},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Qayyum et al. - Unknown - Secure and Robust Machine Learning for Healthcare A Survey.pdf:pdf},
title = {{Secure and Robust Machine Learning for Healthcare: A Survey}},
url = {https://tinyurl.com/FDA-AI-diabetic-eye}
}
@misc{Salay2017,
abstract = {—Machine learning (ML) plays an ever-increasing role in advanced automotive functionality for driver assistance and autonomous operation; however, its adequacy from the perspective of safety certification remains controversial. In this paper, we analyze the impacts that the use of ML as an implementation approach has on ISO 26262 safety lifecycle and ask what could be done to address them. We then provide a set of recommendations on how to adapt the standard to accommodate ML.},
archivePrefix = {arXiv},
arxivId = {1709.02435},
author = {Salay, Rick and Queiroz, Rodrigo and Czarnecki, Krzysztof},
booktitle = {arXiv},
eprint = {1709.02435},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Salay, Queiroz, Czarnecki - 2017 - An Analysis of ISO 26262 Using Machine Learning Safely in Automotive Software.pdf:pdf},
issn = {23318422},
month = {sep},
publisher = {arXiv},
title = {{An Analysis of ISO 26262: Using Machine Learning Safely in Automotive Software}},
year = {2017}
}
@article{Macrae2019a,
abstract = {Artificial intelligence (AI) has enormous potential to improve the safety of healthcare, from increasing diagnostic accuracy,1 to optimising treatment planning,2 to forecasting outcomes of care.3 However, integrating AI technologies into the delivery of healthcare is likely to introduce a range of new risks and amplify existing ones. For instance, failures in widely used software have the potential to quickly affect large numbers of patients4; hidden assumptions in underlying data and models can lead to AI systems delivering dangerous recommendations that are insensitive to local care processes,5 6 and opaque AI techniques such as deep learning can make explaining and learning from failure extremely difficult.7 8 To maximise the benefits of AI in healthcare and to build trust among patients and practitioners, it will therefore be essential to robustly govern the risks that AI poses to patient safety.
In a recent review in this journal, Challen and colleagues present an important and timely analysis of some of the key technological risks associated with the application of machine learning in clinical settings.9 Machine learning is a subfield of AI that focuses on the development of algorithms that are automatically derived and optimised through exposure to large quantities of exemplar ‘training' data.10 The outputs of machine learning algorithms are essentially classifications of patterns that provide some sort of prediction—for instance, predicting whether an image shows a malignant melanoma or a benign mole.11 Some of the basic techniques of machine learning have existed for half a century or more, but progress in the field has accelerated rapidly due to advances in the development of ‘deep' artificial neural networks12 combined with huge increases in computational power and the availability of enormous quantities of data. These techniques have underpinned recent public demonstrations of AI systems {\ldots}},
author = {Macrae, Carl},
doi = {10.1136/bmjqs-2019-009484},
issn = {20445415},
journal = {BMJ Qual. Saf.},
keywords = {governance,information technology,risk management},
month = {jun},
number = {6},
pages = {495--498},
pmid = {30979783},
publisher = {BMJ Publishing Group},
title = {{Governing the safety of artificial intelligence in healthcare}},
url = {https://qualitysafety.bmj.com/content/28/6/495 https://qualitysafety.bmj.com/content/28/6/495.abstract},
volume = {28},
year = {2019}
}
@article{Moin,
abstract = {In this paper, we present the current position of the research project ML-Quadrat, which aims to extend the methodology, modeling language and tool support of ThingML-an open source modeling tool for IoT/CPS-to address Machine Learning needs for the IoT applications. Currently, ThingML offers a modeling language and tool support for modeling the components of the system, their communication interfaces as well as their behaviors. The latter is done through state machines. However, we argue that in many cases IoT/CPS services involve system components and physical processes, whose behaviors are not well understood in order to be modeled using state machines. Hence, quite often a data-driven approach that enables inference based on the observed data, e.g., using Machine Learning is preferred. To this aim, ML-Quadrat integrates the necessary Machine Learning concepts into ThingML both on the modeling level (syntax and semantics of the modeling language) and on the code generators level. We plan to support two target platforms for code generation regarding Stream Processing and Complex Event Processing, namely Apache SAMOA and Apama.},
author = {Moin, Armin and R{\"{o}}ssler, Stephan and G{\"{u}}nnemann, Stephan},
doi = {10.1016/j},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Moin, R{\"{o}}ssler, G{\"{u}}nnemann - Unknown - Augmenting Model-Driven Software Engineering for the Internet of Things with Machine Learning.pdf:pdf},
title = {{Augmenting Model-Driven Software Engineering for the Internet of Things with Machine Learning}},
url = {https://doi.org/10.1016/j.}
}
@article{Bishop2013,
abstract = {Several decades of research in the field of machine learning have resulted in a multitude of different algorithms for solving a broad range of problems. To tackle a new application, a researcher typically tries to map their problem onto one of these existing methods, often influenced by their familiarity with specific algorithms and by the availability of corresponding software implementations. In this study, we describe an alternative methodology for applying machine learning, in which a bespoke solution is formulated for each new application. The solution is expressed through a compact modelling language, and the corresponding custom machine learning code is then generated automatically. This model-based approach offers several major advantages, including the opportunity to create highly tailored models for specific scenarios, as well as rapid prototyping and comparison of a range of alternative models. Furthermore, newcomers to the field of machine learning do not have to learn about the huge range of traditional methods, but instead can focus their attention on understanding a single modelling environment. In this study, we show how probabilistic graphical models, coupled with efficient inference algorithms, provide a very flexible foundation formodel-based machine learning, and we outline a large-scale commercial application of this framework involving tens of millions of users. We also describe the concept of probabilistic programming as a powerful software environment for modelbased machine learning, and we discuss a specific probabilistic programming language called Infer.NET, which has been widely used in practical applications. {\textcopyright} 2012 The Author(s) Published by the Royal Society. All rights reserved.},
author = {Bishop, Christopher M.},
doi = {10.1098/rsta.2012.0222},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Bishop - 2013 - Model-based machine learning.pdf:pdf},
issn = {1364503X},
journal = {Philos. Trans. R. Soc. A Math. Phys. Eng. Sci.},
keywords = {Bayesian inference,Graphical probabilistic programming,Infer. NET},
month = {feb},
number = {1984},
publisher = {Royal Society},
title = {{Model-based machine learning}},
url = {/pmc/articles/PMC3538442/ /pmc/articles/PMC3538442/?report=abstract https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3538442/},
volume = {371},
year = {2013}
}
@article{Kienzle2019,
abstract = {The growing use of models for separating concerns in complex systems has lead to a proliferation of model composition operators. These composition operators have traditionally been defined from scratch following various approaches differing in formality, level of detail, chosen paradigm, and styles. Due to the lack of proper foundations for defining model composition (concepts, abstractions, or frameworks), it is difficult to compare or reuse composition operators. In this paper, we stipulate the existence of a unifying framework that reduces all structural composition operators to structural merging, and all composition operators acting on discrete behaviors to event scheduling. We provide convincing evidence of this hypothesis by discussing how structural and behavioral homogeneous model composition operators (i.e., weavers) can be mapped onto this framework. Based on this discussion, we propose a conceptual model of the framework and identify a set of research challenges, which, if addressed, lead to the realization of this framework to support rigorous and efficient engineering of model composition operators for homogeneous and eventually heterogeneous modeling languages.},
author = {Kienzle, J{\"{o}}rg and Mussbacher, Gunter and Combemale, Benoit and Deantoni, Julien},
doi = {10.1007/s10270-018-00707-8},
issn = {16191374},
journal = {Softw. Syst. Model.},
keywords = {Event scheduling,Event structures,Model composition,Separation of concerns,Symmetric merge},
month = {oct},
number = {5},
pages = {3005--3023},
publisher = {Springer Verlag},
title = {{A unifying framework for homogeneous model composition}},
volume = {18},
year = {2019}
}
@techreport{Alves2018,
author = {Alves, Erin and Bhatt, Devesh and Hall, Brendan and Driscoll, Kevin and Murugesan, Anitha and Rushby, John},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Alves et al. - 2018 - Considerations in Assuring Safety of Increasingly Autonomous Systems Final Technical Report for NASA Project Assur.pdf:pdf},
title = {{Considerations in Assuring Safety of Increasingly Autonomous Systems Final Technical Report for NASA Project Assurance Reasoning for Increasingly Autonomous Systems (ARIAS)}},
year = {2018}
}
@inproceedings{Basciani2016,
abstract = {Over the last years, several model repositories have been proposed in response to the need of the MDE community for advanced systems supporting the reuse of modeling artifacts. Modelers can interact with MDE repositories with different intents ranging from merely repository browsing, to searching specific artifacts satisfying precise requirements. The organization and browsing facilities provided by current repositories is limited since they do not produce structured overviews of the contained artifacts, and the ategorization mechanisms (if any) are based on manual activities. When dealing with large numbers of modeling artifacts, such limitations increase the effort for managing and reusing artifacts stored in model repositories. By focusing on metamodel repositories, in this paper we propose the application of clustering techniques to automatically organize stored metamodels and to provide users with overviews of the application domains covered by the available metamodels. The approach has been implemented in the MDEForge repository.},
author = {Basciani, Francesco and Rocco, Juri Di and Ruscio, Davide Di and Iovino, Ludovico and Pierantonio, Alfonso},
booktitle = {Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics)},
doi = {10.1007/978-3-319-39696-5_21},
isbn = {9783319396958},
issn = {16113349},
keywords = {MDEForge,Metamodel clustering,Model Driven Engineering,Model repositories},
pages = {342--358},
publisher = {Springer Verlag},
title = {{Automated clustering of metamodel repositories}},
volume = {9694},
year = {2016}
}
@misc{Abadi2016Tensor,
abstract = {TensorFlow [1] is an interface for expressing machine learning algorithms, and an implementation for executing such algorithms. A computation expressed using TensorFlow can be executed with little or no change on a wide variety of heterogeneous systems, ranging from mobile devices such as phones and tablets up to large-scale distributed systems of hundreds of machines and thousands of computational devices such as GPU cards. The system is flexible and can be used to express a wide variety of algorithms, including training and inference algorithms for deep neural network models, and it has been used for conducting research and for deploying machine learning systems into production across more than a dozen areas of computer science and other fields, including speech recognition , computer vision, robotics, information retrieval, natural language processing, geographic information extraction, and computational drug discovery. This paper describes the Ten-sorFlow interface and an implementation of that interface that we have built at Google. The TensorFlow API and a reference implementation were released as an open-source package under the Apache 2.0 license in November, 2015 and are available at www.tensorflow.org.},
archivePrefix = {arXiv},
arxivId = {1603.04467v2},
author = {{Mart{\'{i}}n Abadi et al.}},
eprint = {1603.04467v2},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Abadi et al. - Unknown - TensorFlow Large-Scale Machine Learning on Heterogeneous Distributed Systems.pdf:pdf},
howpublished = {www.tensorflow.org},
title = {{TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems}},
url = {www.tensorflow.org},
urldate = {2021-03-17},
year = {2015}
}
@techreport{Manuel2020,
author = {Manuel, Pedro and Costa, Afonso},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Manuel, Costa - 2020 - FACULDADE DE ENGENHARIA DA UNIVERSIDADE DO PORTO Customized Hardware for Long-Short Term Memory Networks in Embed.pdf:pdf},
title = {{FACULDADE DE ENGENHARIA DA UNIVERSIDADE DO PORTO Customized Hardware for Long-Short Term Memory Networks in Embedded Systems}},
year = {2020}
}
@article{Wang2019,
abstract = {Safety is a central dimension in contemporary debates on human health, loss prevention, environmental protection, sustainability, and cleaner production. In the information age, especially in the era of big data, safety information is an essential strategy for safety, and safety informatics has become a major research interest and a popular issue in the field of safety science. In recent years, safety informaticsda new area of safety sciencedhas received increasing attention, developing greatly with successful research on the subject. The three key purposes of this paper are: (i) to analyze the historical development of safety informatics, (ii) to review the research progress of safety informatics, and (iii) to review limitations and propose future directions in the field of safety informatics. First, the development process of safety informatics is divided into four typical stages: (i) the embryonic stage (1940e1980), (ii) the initial stage (1980e1990), (iii) the formation stage (1990e2010), and (iv) the deepening stage (2010epresent). Then, a review of safety informatics research is provided from seven aspects, including: (i) the discipline construction of safety informatics, (ii) theoretical safety information model, (iii) accident causation model from a safety information perspective, (iv) safety management based on safety information, (v) safety big data, (vi) safety intelligence, and (vii) safety information technology. Finally, limitations and future research directions in the safety informatics area are briefly discussed.},
author = {Wang, Bing and Wu, Chao},
doi = {10.1016/j.jclepro.2019.119852},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Wang, Wu - 2019 - Safety informatics as a new, promising and sustainable area of safety science in the information age.pdf:pdf},
keywords = {Information science,Safety 40,Safety informatics,Safety information,Safety science},
title = {{Safety informatics as a new, promising and sustainable area of safety science in the information age}},
url = {https://doi.org/10.1016/j.jclepro.2019.119852},
year = {2019}
}
@techreport{Hernandez-Oralloa,
abstract = {The current analysis in the AI safety literature usually combines a risk or safety issue (e.g., interruptibility) with a particular paradigm for an AI agent (e.g., reinforcement learning). However, there is currently no survey of safety-relevant characteristics of AI systems that may reveal neglected areas of research or suggest to developers what design choices they could make to avoid or minimise certain safety concerns. In this paper, we take a first step towards delivering such a survey , from two angles. The first features AI system characteristics that are already known to be relevant to safety concerns, including internal system characteristics, characteristics relating to the effect of the external environment on the system , and characteristics relating to the effect of the system on the target environment. The second presents a brief survey of a broad range of AI system characteristics that could prove relevant to safety research, including types of interaction, computation, integration, anticipation, supervision, modification , motivation and achievement. This survey enables further work in exploring system characteristics and design choices that affect safety concerns.},
author = {Hern{\'{a}}ndez-Orallo, Jos{\'{e}} and Mart{\'{i}}nez-Plumed, Fernando and Avin, Shahar and Se{\'{a}}n´o, Se{\'{a}}n Se{\'{a}}n´ and Eigeartaigh, ´},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Hern{\'{a}}ndez-Orallo et al. - Unknown - Surveying Safety-relevant AI Characteristics.pdf:pdf},
title = {{Surveying Safety-relevant AI Characteristics}}
}
@article{Fischer1936Iris,
abstract = {When two or more populations have been measured in several characters, special interest attaches to certain functions of the measurements by which the populations are discriminated. At the author's suggestion use has already been made of this fact in craniometry (a) by Mr E. S. Martin, who has applied the principle to the sex differences in measurements of the mandible, and (b) by Miss Mildred Barnard, who showed how to obtain from a series of dated series the particular compound of cranial measurements showing most distinctly a progressive or secular trend. In the present paper the application of the same principle will be illustrated on a taxonomic problem; some questions connected with the precision of the processes employed will also be discussed.},
author = {Fisher, R. A.},
doi = {10.1111/j.1469-1809.1936.tb02137.x},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/FISHER - 1936 - THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMS.pdf:pdf},
issn = {2050-1439},
journal = {Ann. Eugen.},
month = {sep},
number = {2},
pages = {179--188},
publisher = {Wiley},
title = {{The Use of Multiple Measurments in Taxonomic Problems}},
url = {https://onlinelibrary.wiley.com/doi/full/10.1111/j.1469-1809.1936.tb02137.x https://onlinelibrary.wiley.com/doi/abs/10.1111/j.1469-1809.1936.tb02137.x https://onlinelibrary.wiley.com/doi/10.1111/j.1469-1809.1936.tb02137.x},
volume = {7},
year = {1936}
}
@book{taleb2007the,
address = {New York},
author = {Taleb, Nassim},
isbn = {9781400063512},
publisher = {Random House},
title = {{The black swan : the impact of the highly improbable}},
year = {2007}
}
@article{Bencomo2019,
abstract = {More than a decade ago, the research topic models@run.time was coined. Since then, the research area has received increasing attention. Given the prolific results during these years, the current outcomes need to be sorted and classified. Furthermore, many gaps need to be categorized in order to further develop the research topic by experts of the research area but also newcomers. Accordingly, the paper discusses the principles and requirements of models@run.time and the state of the art of the research line. To make the discussion more concrete, a taxonomy is defined and used to compare the main approaches and research outcomes in the area during the last decade and including ancestor research initiatives. We identified and classified 275 papers on models@run.time, which allowed us to identify the underlying research gaps and to elaborate on the corresponding research challenges. Finally, we also facilitate sustainability of the survey over time by offering tool support to add, correct and visualize data.},
author = {Bencomo, Nelly and G{\"{o}}tz, Sebastian and Song, {\textperiodcentered} Hui},
doi = {10.1007/s10270-018-00712-x},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Bencomo, G{\"{o}}tz, Song - 2019 - Models@run.time a guided tour of the state of the art and research challenges.pdf:pdf},
journal = {Softw. Syst. Model.},
keywords = {Causal connection,Models@run.time,Self-reflection,Systematic literature review},
pages = {3049--3082},
title = {{Models@run.time: a guided tour of the state of the art and research challenges}},
url = {https://doi.org/10.1007/s10270-018-00712-x},
volume = {18},
year = {2019}
}
@misc{Challen2019,
abstract = {In medicine, artificial intelligence (AI) research is becoming increasingly focused on applying machine learning (ML) techniques to complex problems, and so allowing computers to make predictions from large amounts of patient data, by learning their own associations.1 Estimates of the impact of AI on the wider economy globally vary wildly, with a recent report suggesting a 14{\%} effect on global gross domestic product by 2030, half of which coming from productivity improvements.2 These predictions create political appetite for the rapid development of the AI industry,3 and healthcare is a priority area where this technology has yet to be exploited.2 3 The digital health revolution described by Duggal et al 4 is already in full swing with the potential to ‘disrupt' healthcare. Health AI research has demonstrated some impressive results,5–10 but its clinical value has not yet been realised, hindered partly by a lack of a clear understanding of how to quantify benefit or ensure patient safety, and increasing concerns about the ethical and medico-legal impact.11
This analysis is written with the dual aim of helping clinical safety professionals to critically appraise current medical AI research from a quality and safety perspective, and supporting research and development in AI by highlighting some of the clinical safety questions that must be considered if medical application of these exciting technologies is to be successful.
Clinical decision support systems (DSS) are in widespread use in medicine and have had most impact providing guidance on the safe prescription of medicines,12 guideline adherence, simple risk screening13 or prognostic scoring.14 These systems use predefined rules, which have predictable behaviour and are usually shown to reduce clinical error,12 although sometimes inadvertently introduce safety issues themselves.15 16 Rules-based systems have also been developed to address diagnostic uncertainty17–19 {\ldots}},
author = {Challen, Robert and Denny, Joshua and Pitt, Martin and Gompels, Luke and Edwards, Tom and Tsaneva-Atanasova, Krasimira},
booktitle = {BMJ Qual. Saf.},
doi = {10.1136/bmjqs-2018-008370},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Challen et al. - 2019 - Artificial intelligence, bias and clinical safety.pdf:pdf;:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Challen et al. - 2019 - Artificial intelligence, bias and clinical safety(2).pdf:pdf},
issn = {20445415},
keywords = {artificial intelligence,clinical decision support systems,clinical safety,machine learning},
month = {mar},
number = {3},
pages = {231--237},
pmid = {30636200},
publisher = {BMJ Publishing Group},
title = {{Artificial intelligence, bias and clinical safety}},
url = {https://qualitysafety.bmj.com/lookup/doi/10.1136/bmjqs-2018-008370 http://qualitysafety.bmj.com/},
volume = {28},
year = {2019}
}
@inproceedings{Burgueno2019,
abstract = {Model-driven engineering (MDE) and Artificial Intelligence (AI) are two separate fields in computer science, which can clearly benefit from cross-fertilization and collaboration. There are at least two ways in which such integrations - which we call MDE Intelligence - can manifest: (1) MDE can benefit from integrating AI concepts and ideas to increasing the power and flexibility of model-driven techniques by means of the application of AI algorithms. (2) Conversely, AI can benefit from integrating concepts and ideas from MDE - for example, using domain-specific languages and model transformations allows domain experts to directly express and manipulate their problems while providing an auditable computation pipeline. To discuss and further stimulate such integrations, the 1st edition of the Workshop on Artificial Intelligence and Model-driven Engineering (MDE Intelligence) was held on September 16, 2019 in Munich, Germany, as part of the satellite events of the IEEE/ACM 22th International Conference on Model-Driven Engineering Languages and Systems (MODELS 2019).},
author = {Burgueno, Loli and Burdusel, Alexandru and Gerard, Sebastien and Wimmer, Manuel},
booktitle = {Proc. - 2019 ACM/IEEE 22nd Int. Conf. Model Driven Eng. Lang. Syst. Companion, Model. 2019},
doi = {10.1109/MODELS-C.2019.00028},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Preface to MDE Intelligence 2019.pdf:pdf},
isbn = {9781728151250},
keywords = {Artificial Intelligence,MDE,MDE Intelligence},
month = {sep},
pages = {168--169},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Preface to MDE intelligence 2019: 1st workshop on artificial intelligence and model-driven engineering}},
year = {2019}
}
@inproceedings{Hartmann2017,
abstract = {Machine learning algorithms are designed to resolve unknown behaviours by extracting commonalities over massive datasets. Unfortunately, learning such global behaviours can be inaccurate and slow for systems composed of heterogeneous elements, which behave very differently, for instance as it is the case for cyber-physical systems and Internet of Things applications. Instead, to make smart decisions, such systems have to continuously refine the behaviour on a per-element basis and compose these small learning units together. However, combining and composing learned behaviours from different elements is challenging and requires domain knowledge. Therefore, there is a need to structure and combine the learned behaviours and domain knowledge together in a flexible way. In this paper we propose to weave machine learning into domain modeling. More specifically, we suggest to decompose machine learning into reusable, chainable, and independently computable small learning units, which we refer to as micro learning units. These micro learning units are modeled together with and at the same level as the domain data. We show, based on a smart grid case study, that our approach can be significantly more accurate than learning a global behaviour while the performance is fast enough to be used for live learning.},
author = {Hartmann, Thomas and Moawad, Assaad and Fouquet, Francois and {Le Traon}, Yves},
booktitle = {Proc. - ACM/IEEE 20th Int. Conf. Model Driven Eng. Lang. Syst. Model. 2017},
doi = {10.1109/MODELS.2017.32},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Hartmann2019{\_}Article{\_}TheNextEvolutionOfMDEASeamless.pdf:pdf},
isbn = {9781538634929},
keywords = {Cyber-physical systems,Domain modeling,Live learning,Meta modeling,Model-driven engineering,Smart grids},
month = {nov},
pages = {180},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{The Next Evolution of MDE: A Seamless Integration of Machine Learning into Domain Modeling}},
year = {2017}
}
@article{Birkenbihl2020,
abstract = {Artificial intelligence (AI) approaches pose a great opportunity for individualized, pre-symptomatic disease diagnosis which plays a key role in the context of personalized, predictive, and finally preventive medicine (PPPM). However, to translate PPPM into clinical practice, it is of utmost importance that AI-based models are carefully validated. The validation process comprises several steps, one of which is testing the model on patient-level data from an independent clinical cohort study. However, recruitment criteria can bias statistical analysis of cohort study data and impede model application beyond the training data. To evaluate whether and how data from independent clinical cohort studies differ from each other, this study systematically compares the datasets collected from two major dementia cohorts, namely, the Alzheimer's Disease Neuroimaging Initiative (ADNI) and AddNeuroMed. The presented comparison was conducted on individual feature level and revealed significant differences among both cohorts. Such systematic deviations can potentially hamper the generalizability of results which were based on a single cohort dataset. Despite identified differences, validation of a previously published, ADNI trained model for prediction of personalized dementia risk scores on 244 AddNeuroMed subjects was successful: External validation resulted in a high prediction performance of above 80{\%} area under receiver operator characteristic curve up to 6 years before dementia diagnosis. Propensity score matching identified a subset of patients from AddNeuroMed, which showed significantly smaller demographic differences to ADNI. For these patients, an even higher prediction performance was achieved, which demonstrates the influence systematic differences between cohorts can have on validation results. In conclusion, this study exposes challenges in external validation of AI models on cohort study data and is one of the rare cases in the neurology field in which such external validation was performed. The presented model represents a proof of concept that reliable models for personalized predictive diagnostics are feasible, which, in turn, could lead to adequate disease prevention and hereby enable the PPPM paradigm in the dementia field.},
author = {Birkenbihl, Colin and Emon, Mohammad Asif and Vrooman, Henri and Westwood, Sarah and Lovestone, Simon and Hofmann-Apitius, Martin and Fr{\"{o}}hlich, Holger},
doi = {10.1007/s13167-020-00216-z},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Birkenbihl et al. - 2020 - Differences in cohort study data affect external validation of artificial intelligence models for predictive.pdf:pdf},
issn = {18785085},
journal = {EPMA J.},
keywords = {Alzheimer's disease,Artificial intelligence,Bioinformatics,Cohort comparison,Cohort data,Data science,Dementia,Digital clinic,Disease modeling,Disease risk prediction,Health data,Individualized patient profiling,Interdisciplinary,Machine learning,Medical data,Model performance,Model validation,Multiprofessional,Neurodegeneration,Precision medicine,Predictive preventive personalized medicine (3 PM/,Propensity score matching,Risk modeling,Sampling bias,Survival analysis,Translational medicine},
month = {sep},
number = {3},
pages = {367--376},
publisher = {Springer Science and Business Media Deutschland GmbH},
title = {{Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice}},
url = {https://pubmed.ncbi.nlm.nih.gov/32843907/},
volume = {11},
year = {2020}
}
@inproceedings{Hoser2019,
abstract = {The implementation of intelligent agents in industrial applications is often prevented by the high cost of adopting such a system to a particular problem domain. This paper states the thesis that when learning agents are applied to work environments that require domain-specific experience, the agent benefits if it can be further adapted by a supervising domain expert. Closely interacting with the agent, a domain expert should be able to understand its decisions and update the underlying knowledge base as needed. The result would be an agent with individualized knowledge that comes in part from the domain experts. The model of such an adaptive learning agent must take into account the problem domain, the design of the learning agent and the perception of the domain user. Therefore, already in the modeling phase, more attention must be paid to make the learning element of the agent adaptable by an operator. Domain modeling and meta-modeling methods could help to make inner processes of the agent more accessible. In addition, the knowledge gained should be made reusable for future agents in similar environments. To begin with, the existing methods for modeling agent systems and the underlying concepts will be evaluated, based on the requirements for different industrial scenarios. The methods are then compiled into a framework that allows for the description and modeling of such systems in terms of adaptability to a problem domain. Where necessary, new methods or tools will be introduced to close the gap between inconsistent modeling artifacts. The framework shall then be used to build learning agents for real-life scenarios and observe their application in a case study. The results will be used to assess the quality of the adapted knowledge base and compare it to a manual knowledge modeling process.},
author = {Hoser, Moritz},
booktitle = {Proc. - 2019 ACM/IEEE 22nd Int. Conf. Model Driven Eng. Lang. Syst. Companion, Model. 2019},
doi = {10.1109/MODELS-C.2019.00101},
isbn = {9781728151250},
keywords = {Adaptive Learning Agents,Domain modeling,Knowledge Engineering,Multi-modeling},
month = {sep},
pages = {660--665},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Modeling adaptive learning agents for domain knowledge transfer}},
year = {2019}
}
@article{Tao2019,
abstract = {With the fast growth of artificial intelligence and big data computing technologies, more and more software service systems have been developed using diverse machine learning models and technologies to make business and intelligent decisions based on their multimedia input to achieve intelligent features, such as image recognition, recommendation, decision making, prediction, etc. Nevertheless, there are increasing quality problems resulting in erroneous testing costs in enterprises and businesses. Existing work seldom discusses how to perform testing and quality validation for AI software. This paper focuses on quality validation for AI software function features. The paper provides our understanding of AI software testing for new features and requirements. In addition, current AI software testing categories are presented and different testing approaches are discussed. Moreover, test quality assessment and criteria analysis are illustrated. Furthermore, a practical study on quality validation for an image recognition system is performed through a metamorphic testing method. Study results show the feasibility and effectiveness of the approach.},
author = {Tao, Chuanqi and Gao, Jerry and Wang, Tiexin},
doi = {10.1109/ACCESS.2019.2937107},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Tao, Gao, Wang - 2019 - Testing and Quality Validation for AI Software-Perspectives, Issues, and Practices.pdf:pdf},
issn = {21693536},
journal = {IEEE Access},
keywords = {AI software quality validation,AI testing,testing AI software},
pages = {120164--120175},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Testing and Quality Validation for AI Software-Perspectives, Issues, and Practices}},
volume = {7},
year = {2019}
}
@book{Piedra2020,
abstract = {Knowledge management plays a crucial role in initiatives that promotes the discovery and sharing of cultural and heritage informational resources from diverse and autonomous organizations and initiatives. Information in this domain is heterogeneous, distributed, multi-lingual, comes in unstructured formats and large quantities, is strongly contextualized by time and place is created collaboratively. These challenges can be undertaken using semantic technologies. Semantic Web approach and Linked Data technologies are a catalyst for cross-domain and cross-organizational semantic data interoperability and data integration. In this regard, this work will look at the interaction between knowledge management and linked open data within the context of cultural and heritage data with a particular focus on the concept of music data. This work will discuss, among other things, how Semantic Knowledge Graphs based on linked data can contribute to enhancing knowledge management and increase productivity, in actions related with cultural data enrichment, data exploration, knowledge discovery, cultural data reuse, and data visualization. By representing information using data models and open semantic standards, data integration and reasoning can be applied to the cultural data in a well-defined way.},
author = {Piedra, Nelson and {Mosquera Ar{\'{e}}valo}, Jean Pa{\'{u}}l},
booktitle = {Adv. Intell. Syst. Comput.},
doi = {10.1007/978-3-030-33547-2_12},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Piedra, Mosquera Ar{\'{e}}valo - 2020 - Linked Data and Musical Information to Improvement the Cultural and Heritage Knowledge Management.pdf:pdf;:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Piedra, Mosquera Ar{\'{e}}valo - 2020 - Linked Data and Musical Information to Improvement the Cultural and Heritage Knowledge Management(2).pdf:pdf},
isbn = {9783030335465},
issn = {21945365},
keywords = {Cultural heritage,Knowledge graphs,Linked open data,Music,Ontologies},
number = {Cimps},
pages = {141--155},
title = {{Linked Data and Musical Information to Improvement the Cultural and Heritage Knowledge Management}},
volume = {1071},
year = {2020}
}
@article{Zhang,
abstract = {Deep learning applications become increasingly popular in important domains such as self-driving systems and facial identity systems. Defective deep learning applications may lead to catastrophic consequences. Although recent research eeorts were made on testing and debugging deep learning applications, the characteristics of deep learning defects have never been studied. To oll this gap, we studied deep learning applications built on top of TensorFlow and collected program bugs related to TensorFlow from StackOver-ow QA pages and Github projects. We extracted information from QA pages, commit messages, pull request messages, and issue discussions to examine the root causes and symptoms of these bugs. We also studied the strategies deployed by TensorFlow users for bug detection and localization. These endings help researchers and TensorFlow users to gain a better understanding of coding defects in TensorFlow programs and point out a new direction for future research. CCS CONCEPTS • Software and its engineering → Software defect analysis; KEYWORDS},
author = {Zhang, Yuhao and Chen, Yifan and Cheung, Shing-Chi and Xiong, Yingfei and Zhang, Lu},
doi = {10.1145/3213846.3213866},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Zhang et al. - Unknown - An Empirical Study on TensorFlow Program Bugs.pdf:pdf},
isbn = {9781450356992},
pages = {12},
publisher = {ACM},
title = {{An Empirical Study on TensorFlow Program Bugs}},
url = {https://doi.org/10.1145/3213846.3213866}
}
@inproceedings{Ribeiro2016,
abstract = {Despite widespread adoption, machine learning models remain mostly black boxes. Understanding the reasons behind predictions is, however, quite important in assessing trust, which is fundamental if one plans to take action based on a prediction, or when choosing whether to deploy a new model. Such understanding also provides insights into the model, which can be used to transform an untrustworthy model or prediction into a trustworthy one. In this work, we propose LIME, a novel explanation technique that explains the predictions of any classifier in an in-terpretable and faithful manner, by learning an interpretable model locally around the prediction. We also propose a method to explain models by presenting representative individual predictions and their explanations in a non-redundant way, framing the task as a submodular optimization problem. We demonstrate the flexibility of these methods by explaining different models for text (e.g. random forests) and image classification (e.g. neural networks). We show the utility of explanations via novel experiments, both simulated and with human subjects, on various scenarios that require trust: deciding if one should trust a prediction, choosing between models, improving an untrustworthy classifier, and identifying why a classifier should not be trusted.},
address = {New York, NY, USA},
author = {Ribeiro, Marco Tulio and Singh, Sameer and Guestrin, Carlos},
booktitle = {Proc. 22nd ACM SIGKDD Int. Conf. Knowl. Discov. Data Min.},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Ribeiro, Singh, Guestrin - Unknown - Why Should I Trust You Explaining the Predictions of Any Classifier.pdf:pdf},
isbn = {9781450342322},
publisher = {ACM},
title = {{"Why Should I Trust You?" Explaining the Predictions of Any Classifier}},
url = {http://dx.doi.org/10.1145/2939672.2939778}
}
@misc{zoo-caffe-21,
howpublished = {http://caffe.berkeleyvision.org/model{\_}zoo.html},
title = {{Caffe | Model Zoo}},
url = {http://caffe.berkeleyvision.org/model{\_}zoo.html},
urldate = {2021-07-19}
}
@article{Adhikari2018,
abstract = {As machine learning models become more accurate, they typically become more
complex and uninterpretable by humans. The black-box character of these models
holds back its acceptance in practice, especially in high-risk domains where
the consequences of failure could be catastrophic such as health-care or
defense. Providing understandable and useful explanations behind ML models or
predictions can increase the trust of the user. Example-based reasoning, which
entails leveraging previous experience with analogous tasks to make a decision,
is a well known strategy for problem solving and justification. This work
presents a new explanation extraction method called LEAFAGE, for a prediction
made by any black-box ML model. The explanation consists of the visualization
of similar examples from the training set and the importance of each feature.
Moreover, these explanations are contrastive which aims to take the
expectations of the user into account. LEAFAGE is evaluated in terms of
fidelity to the underlying black-box model and usefulness to the user. The
results showed that LEAFAGE performs overall better than the current
state-of-the-art method LIME in terms of fidelity, on ML models with non-linear
decision boundary. A user-study was conducted which focused on revealing the
differences between example-based and feature importance-based explanations. It
showed that example-based explanations performed significantly better than
feature importance-based explanation, in terms of perceived transparency,
information sufficiency, competence and confidence. Counter-intuitively, when
the gained knowledge of the participants was tested, it showed that they
learned less about the black-box model after seeing a feature importance-based
explanation than seeing no explanation at all. The participants found feature
importance-based explanation vague and hard to generalize it to other
instances.},
archivePrefix = {arXiv},
arxivId = {1812.09044},
author = {Adhikari, Ajaya and Tax, D. M. J and Satta, Riccardo and Fath, Matthias},
eprint = {1812.09044},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Adhikari et al. - 2018 - LEAFAGE Example-based and Feature importance-based Explanationsfor Black-box ML models.pdf:pdf},
journal = {IEEE Int. Conf. Fuzzy Syst.},
keywords = {eXplainable AI,empirical study,example-based reasoning},
month = {dec},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{LEAFAGE: Example-based and Feature importance-based Explanationsfor Black-box ML models}},
url = {https://arxiv.org/abs/1812.09044v3},
volume = {2019-June},
year = {2018}
}
@techreport{Fulton2018,
abstract = {Formal verification provides a high degree of confidence in safe system operation, but only if reality matches the verified model. Although a good model will be accurate most of the time, even the best models are incomplete. This is especially true in Cyber-Physical Systems because high-fidelity physical models of systems are expensive to develop and often intractable to verify. Conversely, reinforcement learning-based controllers are lauded for their flexibility in unmodeled environments , but do not provide guarantees of safe operation. This paper presents an approach for provably safe learning that provides the best of both worlds: the exploration and optimization capabilities of learning along with the safety guarantees of formal verification. Our main insight is that formal verification combined with verified runtime monitoring can ensure the safety of a learning agent. Verification results are preserved whenever learning agents limit exploration within the confounds of verified control choices as long as observed reality comports with the model used for off-line verification. When a model violation is detected, the agent abandons efficiency and instead attempts to learn a control strategy that guides the agent to a modeled portion of the state space. We prove that our approach toward incorporating knowledge about safe control into learning systems preserves safety guarantees, and demonstrate that we retain the empirical performance benefits provided by reinforcement learning. We also explore various points in the design space for these justified speculative controllers in a simple model of adaptive cruise control model for autonomous cars.},
author = {Fulton, Nathan and Platzer, Andr{\'{e}}},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Fulton, Platzer - 2018 - Safe Reinforcement Learning via Formal Methods Toward Safe Control Through Proof and Learning.pdf:pdf},
keywords = {Robotics Track},
title = {{Safe Reinforcement Learning via Formal Methods: Toward Safe Control Through Proof and Learning}},
url = {www.aaai.org},
year = {2018}
}
@incollection{sutton1992introduction,
author = {Sutton, Richard S},
booktitle = {Reinf. Learn.},
pages = {1--3},
publisher = {Springer},
title = {{Introduction: The challenge of reinforcement learning}},
year = {1992}
}
@article{Kuwajima2020,
abstract = {Fatal accidents are a major issue hindering the wide acceptance of safety-critical systems that employ machine learning and deep learning models, such as automated driving vehicles. In order to use machine learning in a safety-critical system, it is necessary to demonstrate the safety and security of the system through engineering processes. However, thus far, no such widely accepted engineering concepts or frameworks have been established for these systems. The key to using a machine learning model in a deductively engineered system is decomposing the data-driven training of machine learning models into requirement, design, and verification, particularly for machine learning models used in safety-critical systems. Simultaneously, open problems and relevant technical fields are not organized in a manner that enables researchers to select a theme and work on it. In this study, we identify, classify, and explore the open problems in engineering (safety-critical) machine learning systems—that is, in terms of requirement, design, and verification of machine learning models and systems—as well as discuss related works and research directions, using automated driving vehicles as an example. Our results show that machine learning models are characterized by a lack of requirements specification, lack of design specification, lack of interpretability, and lack of robustness. We also perform a gap analysis on a conventional system quality standard SQuaRE with the characteristics of machine learning models to study quality models for machine learning systems. We find that a lack of requirements specification and lack of robustness have the greatest impact on conventional quality models.},
archivePrefix = {arXiv},
arxivId = {1904.00001},
author = {Kuwajima, Hiroshi and Yasuoka, Hirotoshi and Nakae, Toshihiro},
doi = {10.1007/s10994-020-05872-w},
eprint = {1904.00001},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Kuwajima, Yasuoka, Nakae - 2020 - Engineering problems in machine learning systems.pdf:pdf},
issn = {15730565},
journal = {Mach. Learn.},
keywords = {Automated driving,Machine learning,Quality models,Safety critical systems,Software engineering,Systems engineering},
month = {may},
number = {5},
pages = {1103--1126},
publisher = {Springer},
title = {{Engineering problems in machine learning systems}},
volume = {109},
year = {2020}
}
@article{Kuwajima2020a,
abstract = {Fatal accidents are a major issue hindering the wide acceptance of safety-critical systems that employ machine learning and deep learning models, such as automated driving vehicles. In order to use machine learning in a safety-critical system, it is necessary to demonstrate the safety and security of the system through engineering processes. However, thus far, no such widely accepted engineering concepts or frameworks have been established for these systems. The key to using a machine learning model in a deductively engineered system is decomposing the data-driven training of machine learning models into requirement, design, and verification, particularly for machine learning models used in safety-critical systems. Simultaneously, open problems and relevant technical fields are not organized in a manner that enables researchers to select a theme and work on it. In this study, we identify, classify, and explore the open problems in engineering (safety-critical) machine learning systems—that is, in terms of requirement, design, and verification of machine learning models and systems—as well as discuss related works and research directions, using automated driving vehicles as an example. Our results show that machine learning models are characterized by a lack of requirements specification, lack of design specification, lack of interpretability, and lack of robustness. We also perform a gap analysis on a conventional system quality standard SQuaRE with the characteristics of machine learning models to study quality models for machine learning systems. We find that a lack of requirements specification and lack of robustness have the greatest impact on conventional quality models.},
archivePrefix = {arXiv},
arxivId = {1904.00001},
author = {Kuwajima, Hiroshi and Yasuoka, Hirotoshi and Nakae, Toshihiro},
doi = {10.1007/s10994-020-05872-w},
eprint = {1904.00001},
issn = {15730565},
journal = {Mach. Learn.},
keywords = {Automated driving,Machine learning,Quality models,Safety critical systems,Software engineering,Systems engineering},
month = {may},
number = {5},
pages = {1103--1126},
publisher = {Springer},
title = {{Engineering problems in machine learning systems}},
url = {https://doi.org/10.1007/s10994-020-05872-w},
volume = {109},
year = {2020}
}
@article{Combemale2020,
author = {Combemale, Benoit and Kienzle, J{\"{o}}rg and Mussbacher, Gunter and Ali, Hyacinth and Amyot, Daniel and Bagherzadeh, Mojtaba and Batot, Edouard and Bencomo, Nelly and Benni, Benjamin and Bruel, Jean-Michel and Cabot, Jordi and Cheng, Betty H C and Collet, Philippe and Engels, Gregor and Heinrich, Robert and J{\'{e}}z{\'{e}}quel, Jean-Marc and Koziolek, Anne and Reussner, Ralf and Sahraoui, Houari and Saini, Rijul and Sallou, June and Stinckwich, Serge and Syriani, Eugene and Wimmer, Manuel},
doi = {10.1109/MS.2020.2995125ï},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Combemale et al. - 2020 - A Hitchhiker's Guide to Model-Driven Engineering for Data-Centric Systems.pdf:pdf},
pages = {1--9},
title = {{A Hitchhiker's Guide to Model-Driven Engineering for Data-Centric Systems }},
url = {https://hal.inria.fr/hal-02612087},
year = {2020}
}
@inproceedings{Breuker2014,
abstract = {Graphical models and general purpose inference algorithms are powerful tools for moving from imperative towards declarative specification of machine learning problems. Although graphical models define the principle information necessary to adapt inference algorithms to specific probabilistic models, entirely model-driven development is not yet possible. However, generating executable code from graphical models could have several advantages. It could reduce the skills necessary to implement probabilistic models and may speed up development processes. Both advantages address pressing industry needs. They come along with increased supply of data scientist labor, the demand of which cannot be fulfilled at the moment. To explore the opportunities of model-driven big data analytics, I review the main modeling languages used in machine learning as well as inference algorithms and corresponding software implementations. Gaps hampering direct code generation from graphical models are identified and closed by proposing an initial conceptualization of a domain-specific modeling language. {\textcopyright} 2014 IEEE.},
author = {Breuker, Dominic},
booktitle = {Proc. Annu. Hawaii Int. Conf. Syst. Sci.},
doi = {10.1109/HICSS.2014.101},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Towards model-driven engineering for big data analytics–an exploratory analysis of domain- specific languages for machine learning.pdf:pdf},
isbn = {9781479925049},
issn = {15301605},
pages = {758--767},
publisher = {IEEE Computer Society},
title = {{Towards model-driven engineering for big data analytics - An exploratory analysis of domain-specific languages for machine learning}},
year = {2014}
}
@techreport{Deepmind,
abstract = {In many real-world reinforcement learning (RL) problems, besides optimizing the main objective function, an agent must concurrently avoid violating a number of constraints. In particular, besides optimizing performance, it is crucial to guarantee the safety of an agent during training as well as deployment (e.g., a robot should avoid taking actions-exploratory or not-which irrevocably harm its hardware). To incorporate safety in RL, we derive algorithms under the framework of constrained Markov decision processes (CMDPs), an extension of the standard Markov decision processes (MDPs) augmented with constraints on expected cumulative costs. Our approach hinges on a novel Lyapunov method. We define and present a method for constructing Lyapunov functions, which provide an effective way to guarantee the global safety of a behavior policy during training via a set of local linear constraints. Leveraging these theoretical underpinnings, we show how to use the Lyapunov approach to systematically transform dynamic programming (DP) and RL algorithms into their safe counterparts. To illustrate their effectiveness, we evaluate these algorithms in several CMDP planning and decision-making tasks on a safety benchmark domain. Our results show that our proposed method significantly outperforms existing baselines in balancing constraint satisfaction and performance.},
author = {Deepmind, Yinlam Chow and Nachum, Ofir and Brain, Google and Ghavamzadeh, Mohammad and Duenez, Edgar and Deepmind, Guzman},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Deepmind et al. - Unknown - A Lyapunov-based Approach to Safe Reinforcement Learning.pdf:pdf},
title = {{A Lyapunov-based Approach to Safe Reinforcement Learning}}
}
@article{Gu2019,
abstract = {Deep learning-based techniques have achieved state-of-the-art performance on
a wide variety of recognition and classification tasks. However, these networks
are typically computationally expensive to train, requiring weeks of
computation on many GPUs; as a result, many users outsource the training
procedure to the cloud or rely on pre-trained models that are then fine-tuned
for a specific task. In this paper we show that outsourced training introduces
new security risks: an adversary can create a maliciously trained network (a
backdoored neural network, or a $\backslash$emph{\{}BadNet{\}}) that has state-of-the-art
performance on the user's training and validation samples, but behaves badly on
specific attacker-chosen inputs. We first explore the properties of BadNets in
a toy example, by creating a backdoored handwritten digit classifier. Next, we
demonstrate backdoors in a more realistic scenario by creating a U.S. street
sign classifier that identifies stop signs as speed limits when a special
sticker is added to the stop sign; we then show in addition that the backdoor
in our US street sign detector can persist even if the network is later
retrained for another task and cause a drop in accuracy of {\{}25{\}}$\backslash${\%} on average
when the backdoor trigger is present. These results demonstrate that backdoors
in neural networks are both powerful and---because the behavior of neural
networks is difficult to explicate---stealthy. This work provides motivation
for further research into techniques for verifying and inspecting neural
networks, just as we have developed tools for verifying and debugging software.},
archivePrefix = {arXiv},
arxivId = {1708.06733},
author = {Gu, Tianyu and Dolan-Gavitt, Brendan and Garg, Siddharth},
eprint = {1708.06733},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Gu, Dolan-Gavitt, Garg - 2017 - BadNets Identifying Vulnerabilities in the Machine Learning Model Supply Chain.pdf:pdf},
month = {aug},
title = {{BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain}},
url = {https://arxiv.org/abs/1708.06733v2},
year = {2017}
}
@inproceedings{Levinson2011,
abstract = {In order to achieve autonomous operation of a vehicle in urban situations with unpredictable traffic, several realtime systems must interoperate, including environment perception, localization, planning, and control. In addition, a robust vehicle platform with appropriate sensors, computational hardware, networking, and software infrastructure is essential. {\textcopyright} 2011 IEEE.},
author = {Levinson, Jesse and Askeland, Jake and Becker, Jan and Dolson, Jennifer and Held, David and Kammel, Soeren and Kolter, J. Zico and Langer, Dirk and Pink, Oliver and Pratt, Vaughan and Sokolsky, Michael and Stanek, Ganymed and Stavens, David and Teichman, Alex and Werling, Moritz and Thrun, Sebastian},
booktitle = {IEEE Intell. Veh. Symp. Proc.},
doi = {10.1109/IVS.2011.5940562},
pages = {163--168},
title = {{Towards fully autonomous driving: Systems and algorithms}},
year = {2011}
}
@inproceedings{Nguyen2019,
abstract = {Manual classification methods of metamodel repositories require highly trained personnel and the results are usually influenced by the subjectivity of human perception. Therefore, automated metamodel classification is very desirable and stringent. In this work, Machine Learning techniques have been employed for metamodel automated classification. In particular, a tool implementing a feed-forward neural network is introduced to classify metamodels. An experimental evaluation over a dataset of 555 metamodels demonstrates that the technique permits to learn from manually classified data and effectively categorize incoming unlabeled data with a considerably high prediction rate: the best performance comprehends 95.40{\%} as success rate, 0.945 as precision, 0.938 as recall, and 0.942 as F1 score.},
author = {Nguyen, Phuong T. and {Di Rocco}, Juri and {Di Ruscio}, Davide and Pierantonio, Alfonso and Iovino, Ludovico},
booktitle = {Proc. - 2019 ACM/IEEE 22nd Int. Conf. Model Driven Eng. Lang. Syst. Model. 2019},
doi = {10.1109/MODELS.2019.00011},
isbn = {9781728125350},
keywords = {Machine learning,metamodel classification,metamodel repositories},
month = {sep},
pages = {272--282},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Automated Classification of Metamodel Repositories: A Machine Learning Approach}},
year = {2019}
}
@article{iso15026,
doi = {10.1109/IEEESTD.2014.6948215},
journal = {IEEE Std 15026-1-2014},
pages = {1--45},
title = {{IEEE Standard Adoption of ISO/IEC 15026-1--Systems and Software Engineering--Systems and Software Assurance--Part 1: Concepts and Vocabulary}},
year = {2014}
}
@article{Egert2020,
abstract = {Artificial intelligence (AI) and machine learning (ML) have the potential to improve multiple facets of medical practice, including diagnosis of disease, surgical training, clinical outcomes, and access to healthcare. There have been various applications of this technology to surgical fields. AI and ML have been used to evaluate a surgeon's technical skill. These technologies can detect instrument motion, recognize patterns in video recordings, and track the physical motion, eye movements, and cognitive function of the surgeon. These modalities also aid in the advancement of robotic surgical training. The da Vinci Standard Surgical System developed a recording and playback system to help trainees receive tactical feedback to acquire more precision when operating. ML has shown promise in recognizing and classifying complex patterns on diagnostic images and within pathologic tissue analysis. This allows for more accurate and efficient diagnosis and treatment. Artificial neural networks are able to analyze sets of symptoms in conjunction with labs, imaging, and exam findings to determine the likelihood of a diagnosis or outcome. Telemedicine is another use of ML and AI that uses technology such as voice recognition to deliver health care remotely. Limitations include the need for large data sets to program computers to create the algorithms. There is also the potential for misclassification of data points that do not follow the typical patterns learned by the machine. As more applications of AI and ML are developed for the surgical field, further studies are needed to determine feasibility, efficacy, and cost.},
author = {Egert, Melissa and Steward, James E. and Sundaram, Chandru P.},
doi = {10.1007/S13193-020-01166-8},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Egert, Steward, Sundaram - 2020 - Machine Learning and Artificial Intelligence in Surgical Fields.pdf:pdf},
issn = {0976-6952},
journal = {Indian J. Surg. Oncol. 2020 114},
keywords = {Oncology,Surgery,Surgical Oncology},
month = {jul},
number = {4},
pages = {573--577},
publisher = {Springer},
title = {{Machine Learning and Artificial Intelligence in Surgical Fields}},
url = {https://link.springer.com/article/10.1007/s13193-020-01166-8},
volume = {11},
year = {2020}
}
@article{Henriksson2018,
abstract = {Machine learning (ML) applications generate a continuous stream of success stories from various domains. ML enables many novel applications, also in safety-critical contexts. However, the functional safety standards such as ISO 26262 did not evolve to cover ML. We conduct an exploratory study on which parts of ISO 26262 represent the most critical gaps between safety engineering and ML development. While this paper only reports the first steps toward a larger research endeavor, we report three adaptations that are critically needed to allow ISO 26262 compliant engineering, and related suggestions on how to evolve the standard. CCS CONCEPTS • Computing methodologies → Machine learning; • Software and its engineering → Software safety; KEYWORDS automotive software, machine learning, safety, interview study ACM Reference Format: Jens Henriksson, Markus Borg, and Cristofer Englund. 2018. Automotive safety and machine learning: Initial results from a study on how to adapt the ISO 26262 safety standard.},
archivePrefix = {arXiv},
arxivId = {1604.07316},
author = {Henriksson, Jens and Borg, Markus and Englund, Cristofer},
doi = {10.1145/3194085.3194090},
eprint = {1604.07316},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Henriksson, Borg, Englund - 2018 - Automotive safety and machine learning Initial results from a study on how to adapt the ISO 26262 saf.pdf:pdf},
isbn = {9781450357395},
publisher = {ACM},
title = {{Automotive safety and machine learning: Initial results from a study on how to adapt the ISO 26262 safety standard}},
url = {https://doi.org/10.1145/3194085.3194090},
year = {2018}
}
@article{Lessmann2015,
author = {Lessmann, Stefan and Baesens, Bart and Seow, Hsin-Vonn and Thomas, Lyn C},
doi = {10.1016/j.ejor.2015.05.030},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Lessmann et al. - 2015 - Benchmarking state-of-the-art classification algorithms for credit scoring An update of research.pdf:pdf},
journal = {Eur. J. Oper. Res.},
keywords = {Credit scoring,Data mining,Forecasting benchmark,OR in banking},
pages = {124--136},
title = {{Benchmarking state-of-the-art classification algorithms for credit scoring: An update of research}},
url = {http://dx.doi.org/10.1016/j.ejor.2015.05.030},
volume = {247},
year = {2015}
}
@article{Mahendran2015,
abstract = {Image representations, from SIFT and Bag of Visual Words to Convolutional
Neural Networks (CNNs), are a crucial component of almost any image
understanding system. Nevertheless, our understanding of them remains limited.
In this paper we conduct a direct analysis of the visual information contained
in representations by asking the following question: given an encoding of an
image, to which extent is it possible to reconstruct the image itself? To
answer this question we contribute a general framework to invert
representations. We show that this method can invert representations such as
HOG and SIFT more accurately than recent alternatives while being applicable to
CNNs too. We then use this technique to study the inverse of recent
state-of-the-art CNN image representations for the first time. Among our
findings, we show that several layers in CNNs retain photographically accurate
information about the image, with different degrees of geometric and
photometric invariance.},
archivePrefix = {arXiv},
arxivId = {1412.0035},
author = {Mahendran, Aravindh and Vedaldi, Andrea},
eprint = {1412.0035},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Mahendran, Vedaldi - 2014 - Understanding Deep Image Representations by Inverting Them.pdf:pdf},
journal = {Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit.},
month = {nov},
pages = {5188--5196},
publisher = {IEEE Computer Society},
title = {{Understanding Deep Image Representations by Inverting Them}},
url = {https://arxiv.org/abs/1412.0035v1},
volume = {07-12-June-2015},
year = {2014}
}
@inproceedings{Hartmann2019,
abstract = {Although artificial intelligence and machine learning are currently extremely fashionable, applying machine learning on real-life problems remains very challenging. Data scientists need to evaluate various learning algorithms and tune their numerous parameters, based on their assumptions and experience, against concrete problems and training data sets. This is a long, tedious, and resource expensive task. Meta-learning is a recent technique to overcome, i.e. automate this problem. It aims at using machine learning itself to automatically learn the most appropriate algorithms and parameters for a machine learning problem. As it turns out, there are many parallels between meta-modelling - in the sense of model-driven engineering - and meta-learning. Both rely on abstractions, the meta data, to model a predefined class of problems and to define the variabilities of the models conforming to this definition. Both are used to define the output and input relationships and then fitting the right models to represent that behaviour. In this paper, we envision how a meta-model for meta-learning can look like. We discuss possible variabilities, for what types of learning it could be appropriate for, how concrete learning models can be generated from it, and how models can be finally selected. Last but not least, we discuss a possible integration into existing modelling tools.},
author = {Hartmann, Thomas and Moawad, Assaad and Schockaert, Cedric and Fouquet, Francois and {Le Traon}, Yves},
booktitle = {Proc. - 2019 ACM/IEEE 22nd Int. Conf. Model Driven Eng. Lang. Syst. Model. 2019},
doi = {10.1109/MODELS.2019.00014},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Meta-Modelling Meta-Learning.pdf:pdf},
isbn = {9781728125350},
keywords = {AutoML,meta-learning,meta-modelling,modelling framework},
month = {sep},
pages = {300--305},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Meta-Modelling Meta-Learning}},
year = {2019}
}
@techreport{EllDocu,
abstract = {We present VERIFAI, a software toolkit for the formal design and analysis of systems that include artificial intelligence (AI) and machine learning (ML) components. VERIFAI particularly addresses challenges with applying formal methods to ML components such as perception systems based on deep neural networks, as well as systems containing them, and to model and analyze system behavior in the presence of environment uncertainty. We describe the initial version of VERIFAI, which centers on simulation-based verification and synthesis, guided by formal models and specifications. We give examples of several use cases, including temporal-logic falsification, model-based systematic fuzz testing , parameter synthesis, counterexample analysis, and data set augmentation.},
author = {{Ell Docu}, W and Dreossi, Tommaso and Fremont, Daniel J and Ghosh, Shromona and Kim, Edward and Ravanbakhsh, Hadi and Vazquez-Chanlatte, Marcell and Seshia, Sanjit A},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Ell Docu et al. - Unknown - Con sis te n t Comple te VERIFAI A Toolkit for the Formal Design and Analysis of Artificial Intelligence-Ba.pdf:pdf},
keywords = {Artificial,Autonomous,Cyber-,Falsification {\textperiodcentered},Formal,Intelligence {\textperiodcentered},Learning {\textperiodcentered},Machine,Methods {\textperiodcentered},Physical,Simulation {\textperiodcentered},Sys-tems {\textperiodcentered},Vehicles},
title = {{Con sis te n t * Comple te VERIFAI: A Toolkit for the Formal Design and Analysis of Artificial Intelligence-Based Systems}}
}
@article{Ashmore2018,
abstract = {Training data has a significant influence on the behaviour of an artificial intelligence algorithm developed using machine learning techniques. Consequently, any argument that the trained algorithm is, in some way, fit for purpose ought to include consideration of data as an entity in its own right. We describe some simple techniques that can provide domain experts and algorithm developers with insights into training data and which can be implemented without specialist computer hardware. Specifically, we consider sampling density, test case generation and monitoring for distribution shift. The techniques are illustrated using example data sets from the University of California, Irvine, Machine Learning repository.},
author = {Ashmore, Rob and Hill, Matthew},
doi = {10.1007/978-3-319-99229-7_33},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Ashmore, Hill - 2018 - “Boxing Clever” Practical Techniques for Gaining Insights into Training Data and Monitoring Distribution Shift.pdf:pdf},
journal = {Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics)},
keywords = {Artificial intelligence,Distribution shift,Machine learning,Test cases,Training data},
month = {sep},
pages = {393--405},
publisher = {Springer, Cham},
title = {{“Boxing Clever”: Practical Techniques for Gaining Insights into Training Data and Monitoring Distribution Shift}},
url = {https://link.springer.com/chapter/10.1007/978-3-319-99229-7{\_}33},
volume = {11094 LNCS},
year = {2018}
}
@article{Waa2018,
abstract = {End-users of machine learning-based systems benefit from measures that quantify the trustworthiness of the underlying models. Measures like accuracy provide for a general sense of model performance, but offer no detailed information on specific model outputs. Probabilistic outputs, on the other hand, express such details, but they are not available for all types of machine learning, and can be heavily influenced by bias and lack of representative training data. Further, they are often difficult to understand for non-experts. This study proposes an intuitive certainty measure (ICM) that produces an accurate estimate of how certain a machine learning model is for a specific output, based on errors it made in the past. It is designed to be easily explainable to non-experts and to act in a predictable, reproducible way. ICM was tested on four synthetic tasks solved by support vector machines, and a real-world task solved by a deep neural network. Our results show that ICM is both more accurate and intuitive than related approaches. Moreover, ICM is neutral with respect to the chosen machine learning model, making it widely applicable.},
author = {{Van Der Waa}, Jasper and {Van Diggelen}, Jurriaan and Neerincx, Mark and Raaijmakers, Stephan},
doi = {10.5220/0006542603140321},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Van Der Waa et al. - 2018 - ICM An Intuitive Model Independent and Accurate Certainty Measure for Machine Learning(2).pdf:pdf},
isbn = {9789897582752},
keywords = {Certainty,Explainable Artificial Intelligence,Machine learning,Trust,Uncertainty},
title = {{ICM: An Intuitive Model Independent and Accurate Certainty Measure for Machine Learning}},
year = {2018}
}
@article{Morikawa2020,
abstract = {In recent years, curial incidents and accidents have been reported due to un-intended control caused by misjudgment of statistical machine learning (SML), which include deep learning. The international functional safety standards for Electric/Electronic/Programmable (E/E/P) systems have been widely spread to improve safety. However, most of them do not recom-mended to use SML in safety critical systems so far. In practical the new concepts and methods are urgently required to enable SML to be safely used in safety critical systems. In this paper, we organize five kinds of technical safety concepts (TSCs) for SML components toward accordance with functional safety standards. We discuss not only quantitative evaluation criteria, but also development process based on XAI (eXplainable Artificial Intelligence) and Automotive SPICE to improve explainability and reliability in development phase. Fi-nally, we briefly compare the TSCs in cost and difficulty, and expect to en-courage further discussion in many communities and domain.},
archivePrefix = {arXiv},
arxivId = {2008.01263},
author = {Morikawa, Akihisa and Matsubara, Yutaka},
eprint = {2008.01263},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Morikawa, Matsubara - 2020 - Safety design concepts for statistical machine learning components toward accordance with functional safety.pdf:pdf},
keywords = {Functional safety,Safety design concept,Statistical machine learning},
month = {aug},
title = {{Safety design concepts for statistical machine learning components toward accordance with functional safety standards}},
url = {http://arxiv.org/abs/2008.01263},
year = {2020}
}
@techreport{Hernandez-Orallo,
abstract = {AI safety often analyses a risk or safety issue, such as interruptibility, under a particular AI paradigm, such as reinforcement learning. But what is an AI paradigm and how does it affect the understanding and implications of the safety issue? Is AI safety research covering the most representative paradigms and the right combinations of paradigms with safety issues? Will current research directions in AI safety be able to anticipate more capable and powerful systems yet to come? In this paper we analyse these questions, introducing a distinction between two types of paradigms in AI: artefacts and techniques. We then use experimental data of research and media documents from AI Topics, an official publication of the AAAI, to examine how safety research is distributed across artefacts and techniques. We observe that AI safety research is not sufficiently antici-patory, and is heavily weighted towards certain research paradigms. We identify a need for AI safety to be more explicit about the artefacts and techniques for which a particular issue may be applicable, in order to identify gaps and cover a broader range of issues.},
author = {Hern{\'{a}}ndez-Orallo, Jose and Mart{\'{i}}nez-Plumed, Fernando and Avin, Shahar and Whittlestone, Jess and Se{\'{a}}n´o, Se{\'{a}}n Se{\'{a}}n´ and Eigeartaigh, ´},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Hern{\'{a}}ndez-Orallo et al. - Unknown - AI Paradigms and AI Safety Mapping Artefacts and Techniques to Safety Issues.pdf:pdf},
title = {{AI Paradigms and AI Safety: Mapping Artefacts and Techniques to Safety Issues}},
url = {https://github.com/nandomp/AIParadigmsSafety.}
}
@article{Zolotas2019,
abstract = {Flexible or bottom-up model-driven engineering (MDE) is an emerging approach to domain and systems modelling. Domain experts, who have detailed domain knowledge, typically lack the technical expertise to transfer this knowledge using traditional MDE tools. Flexible MDE approaches tackle this challenge by promoting the use of simple drawing tools to increase the involvement of domain experts in the language definition process. In such approaches, no metamodel is created upfront, but instead the process starts with the definition of example models that will be used to infer the metamodel. Pre-defined metamodels created by MDE experts may miss important concepts of the domain and thus restrict their expressiveness. However, the lack of a metamodel, that encodes the semantics of conforming models has some drawbacks, among others that of having models with elements that are unintentionally left untyped. In this paper, we propose the use of classification algorithms to help with the inference of such untyped elements. We evaluate the proposed approach in a number of random generated example models from various domains. The correct type prediction varies from 23 to 100{\%} depending on the domain, the proportion of elements that were left untyped and the prediction algorithm used. Keywords Model-driven engineering {\textperiodcentered} Flexible model-driven engineering {\textperiodcentered} Bottom-up metamodelling {\textperiodcentered} Type inference {\textperiodcentered} Classification and regression trees {\textperiodcentered} Random forests},
author = {Zolotas, Athanasios and Matragkas, Nicholas and Devlin, {\textperiodcentered} Sam and Kolovos, Dimitrios S and Richard, {\textperiodcentered} and Paige, F},
doi = {10.1007/s10270-018-0658-5},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Zolotas et al. - 2019 - Type inference in flexible model-driven engineering using classification algorithms.pdf:pdf},
journal = {Softw. Syst. Model.},
keywords = {Bottom-up metamodelling,Classification and regression trees,Flexible model-driven engineering,Model-driven engineering,Random forests,Type inference},
pages = {345--366},
title = {{Type inference in flexible model-driven engineering using classification algorithms}},
url = {https://doi.org/10.1007/s10270-018-0658-5},
volume = {18},
year = {2019}
}
@book{mitchell1997machine,
author = {Mitchell, T M},
isbn = {9780070428072},
publisher = {McGraw-Hill Education},
series = {McGraw-Hill international editions - computer science series},
title = {{Machine Learning}},
url = {https://books.google.ca/books?id=xOGAngEACAAJ},
year = {1997}
}
@techreport{koopman2016challenges,
abstract = {Software testing is all too often simply a bug hunt rather than a well-considered exercise in ensuring quality. A more methodical approach than a simple cycle of system-level test-fail-patch-test will be required to deploy safe autonomous vehicles at scale. The ISO 26262 development V process sets up a framework that ties each type of testing to a corresponding design or requirement document, but presents challenges when adapted to deal with the sorts of novel testing problems that face autonomous vehicles. This paper identifies five major challenge areas in testing according to the V model for autonomous vehicles: driver out of the loop, complex requirements, non-deterministic algorithms, inductive learning algorithms, and fail-operational systems. General solution approaches that seem promising across these different challenge areas include: phased deployment using successively relaxed operational scenarios, use of a monitor/actuator pair architecture to separate the most complex autonomy functions from simpler safety functions, and fault injection as a way to perform more efficient edge case testing. While significant challenges remain in safety-certifying the type of algorithms that provide high-level autonomy themselves, it seems within reach to instead architect the system and its accompanying design process to be able to employ existing software safety approaches.},
author = {Koopman, Philip and Wagner, Michael},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd/Mendeley Desktop/Downloaded/Challenges in Autonomous Vehicle testing and validation.pdf:pdf},
pages = {15--24},
title = {{Challenges in Autonomous Vehicle Testing and Validation}},
year = {2016}
}
@techreport{Foster2014,
abstract = {A large number of papers are appearing in the biomedical engineering literature that describe the use of machine learning techniques to develop classifiers for detection or diagnosis of disease. However, the usefulness of this approach in developing clinically validated diagnostic techniques so far has been limited and the methods are prone to overfitting and other problems which may not be immediately apparent to the investigators. This commentary is intended to help sensitize investigators as well as readers and reviewers of papers to some potential pitfalls in the development of classifiers, and suggests steps that researchers can take to help avoid these problems. Building classifiers should be viewed not simply as an add-on statistical analysis, but as part and parcel of the experimental process. Validation of classifiers for diagnostic applications should be considered as part of a much larger process of establishing the clinical validity of the diagnostic technique.},
author = {Foster, Kenneth R and Koprowski, Robert and Skufca, Joseph D},
doi = {10.1186/1475-925X-13-94},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Foster, Koprowski, Skufca - 2014 - Machine learning, medical diagnosis, and biomedical engineering research-commentary.pdf:pdf},
keywords = {Artificial intelligence,Classifiers,Image processing,Machine learning,Support vector machine},
title = {{Machine learning, medical diagnosis, and biomedical engineering research-commentary}},
url = {http://www.biomedical-engineering-online.com/content/13/1/94},
year = {2014}
}
@book{bishop2006pattern,
author = {Bishop, C M},
file = {:C$\backslash$:/Users/Suchendra/OneDrive - McMaster University/PhD Thesis/Books/Bishop - Pattern Recognition And Machine Learning - Springer 2006.pdf:pdf},
isbn = {9780387310732},
publisher = {Springer},
series = {Information Science and Statistics},
title = {{Pattern Recognition and Machine Learning}},
url = {https://books.google.ca/books?id=qWPwnQEACAAJ},
year = {2006}
}
@article{Aydin2019,
abstract = {Drones are currently being used for commercial, public safety, hobby, and scientific research purposes. Short term and long term forecasts show extensive utilization of drones in various sectors. Yet, it is uncertain if general public accepts this highly potential drone congestion. This quantitative survey study explored public acceptance of drones by using the Knowledge, Attitude and Practice (KAP) model and using statistical analysis to decrease the uncertainty. The KAP model was utilized in a recent survey study to understand the public opinion of drones. This survey instrument was thoroughly redesigned to better understand general public's knowledge and attitudes of forty different drone applications, and associated risk perceptions. The practice section was redesigned as well to better understand how public uses drones personally and on company level. The results showed that drones were not well accepted at present except for public safety and scientific research applications. Commercial and hobby uses are not supported. Public sees the drones as a risky technology that directly interferes with their privacy. Moreover, public is not aware of most of the future drone applications, and many current applications. The survey was distributed to both general public and stakeholders to represent both part of the society (153 fully completed surveys). A qualitative risk analysis was conducted in this study as well. By considering the results, this study recommends that public and private institutions should collaborate to develop risk mitigation and response strategies to minimize the risks. Also, public needs to get informed on these strategies through mass media and educational institutions. This might help improving the reputation of drones from killer machines or privacy interrupters towards a technology that helps our society.},
author = {Aydin, Burchan},
doi = {10.1016/J.TECHSOC.2019.101180},
issn = {0160-791X},
journal = {Technol. Soc.},
month = {nov},
pages = {101180},
publisher = {Pergamon},
title = {{Public acceptance of drones: Knowledge, attitudes, and practice}},
volume = {59},
year = {2019}
}
@article{Shlezinger2020,
abstract = {Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques. Such model-based methods utilize mathematical formulations that represent the underlying physics, prior information and additional domain knowledge. Simple classical models are useful but sensitive to inaccuracies and may lead to poor performance when real systems display complex or dynamic behavior. On the other hand, purely data-driven approaches that are model-agnostic are becoming increasingly popular as datasets become abundant and the power of modern deep learning pipelines increases. Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance, especially for supervised problems. However, DNNs typically require massive amounts of data and immense computational resources, limiting their applicability for some signal processing scenarios. We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches. Such model-based deep learning methods exploit both partial domain knowledge, via mathematical structures designed for specific problems, as well as learning from limited data. In this article we survey the leading approaches for studying and designing model-based deep learning systems. We divide hybrid model-based/data-driven systems into categories based on their inference mechanism. We provide a comprehensive review of the leading approaches for combining model-based algorithms with deep learning in a systematic manner, along with concrete guidelines and detailed signal processing oriented examples from recent literature. Our aim is to facilitate the design and study of future systems on the intersection of signal processing and machine learning that incorporate the advantages of both domains.},
archivePrefix = {arXiv},
arxivId = {2012.08405},
author = {Shlezinger, Nir and Whang, Jay and Eldar, Yonina C. and Dimakis, Alexandros G.},
eprint = {2012.08405},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Shlezinger et al. - 2020 - Model-Based Deep Learning.pdf:pdf},
journal = {arXiv},
month = {dec},
publisher = {arXiv},
title = {{Model-Based Deep Learning}},
url = {http://arxiv.org/abs/2012.08405},
year = {2020}
}
@article{iot_apps_conf,
abstract = {With the advent of rapid developments, large number of heterogeneous devices is able to connect with the help of IOT technology. Although IOT possess very complex architecture because of connectivity of variety of devices and services in the system. In this paper, a brief concept of urban IOT system is presented which are designed to support smart city and advanced communication technologies. Hence a comprehensive survey of architecture, technologies, and computational frameworks is provided for a smart IOT. It also discusses the major vulnerabilities and challenges faced by IOT and also present how machine learning is applied to IOT. Hence smart city is considered as the use case and it explains how various techniques are applied to data in order to extract great results with good efficiency.},
author = {Jindal, Mansi and Gupta, Jatin and Bhushan, Bharat},
doi = {10.1109/ICCCIS48478.2019.8974551},
journal = {Proc. - 2019 Int. Conf. Comput. Commun. Intell. Syst. ICCCIS 2019},
keywords = {Internet of Things,Machine Learning,knn (k nearest neighbors),naive bayes,supervised learning,support vector machines},
month = {oct},
pages = {430--434},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
title = {{Machine learning methods for IoT and their Future Applications}},
volume = {2019-January},
year = {2019}
}
@article{Bishnu2015,
abstract = {Teramoto et al. defined a new measure called the gap ratio that measures the
uniformity of a finite point set sampled from {\$}\backslashcal S{\$}, a bounded subset of
{\$}\backslashmathbb{\{}R{\}}{\^{}}2{\$}. We generalize this definition of measure over all metric spaces
by appealing to covering and packing radius. The definition of gap ratio needs
only a metric unlike discrepancy, a widely used uniformity measure, that
depends on the notion of a range space and its volume. We also show some
interesting connections of gap ratio to Delaunay triangulation and discrepancy
in the Euclidean plane. The major focus of this work is on solving optimization
related questions about selecting uniform point samples from metric spaces; the
uniformity being measured using gap ratio. We consider discrete spaces like
graph and set of points in the Euclidean space and continuous spaces like the
unit square and path connected spaces. We deduce lower bounds, prove hardness
and approximation hardness results. We show that a general approximation
algorithm framework gives different approximation ratios for different metric
spaces based on the lower bound we deduce. Apart from the above, we show
existence of coresets for sampling uniform points from the Euclidean space --
for both the static and the streaming case. This leads to a {\$}\backslashleft( 1+\backslashepsilon
\backslashright){\$}-approximation algorithm for uniform sampling from the Euclidean space.},
archivePrefix = {arXiv},
arxivId = {1411.7819},
author = {Bishnu, Arijit and Desai, Sameer and Ghosh, Arijit and Goswami, Mayank and Paul, Subhabrata},
eprint = {1411.7819},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Bishnu et al. - 2014 - Uniformity of point samples in metric spaces using gap ratio.pdf:pdf},
journal = {SIAM J. Discret. Math.},
keywords = {Approximation,Gap ratio,Hardness,Lower bounds,Metric space,Uniformity measure},
month = {nov},
number = {3},
pages = {2138--2171},
publisher = {Society for Industrial and Applied Mathematics Publications},
title = {{Uniformity of point samples in metric spaces using gap ratio}},
url = {https://arxiv.org/abs/1411.7819v2},
volume = {31},
year = {2014}
}
@article{Burton,
abstract = {This paper describes the challenges involved in arguing the safety of highly automated driving functions which make use of machine learning techniques. An assurance case structure is used to highlight the systems engineering and validation considerations when applying machine learning methods for highly automated driving. Particular focus is placed on addressing functional insufficiencies in the perception functions based on convolutional neural networks and possible types of evidence that can be used to mitigate against such risks.},
author = {Burton, Simon and Gauerhof, Lydia and Heinzemann, Christian},
doi = {10.1007/978-3-319-66284-8},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Burton, Gauerhof, Heinzemann - Unknown - Making the Case for Safety of Machine Learning in Highly Automated Driving.pdf:pdf},
title = {{Making the Case for Safety of Machine Learning in Highly Automated Driving}},
url = {https://www.youtube.com/watch?v=u6aEYuemt0M}
}
@article{Macrae2019,
author = {Macrae, Carl},
doi = {10.1136/bmjqs-2019-009484},
issn = {20445415},
journal = {BMJ Qual. Saf.},
keywords = {governance,information technology,risk management},
month = {jun},
number = {6},
pages = {495--498},
pmid = {30979783},
publisher = {BMJ Publishing Group},
title = {{Governing the safety of artificial intelligence in healthcare}},
url = {https://qualitysafety.bmj.com/lookup/doi/10.1136/bmjqs-2019-009484},
volume = {28},
year = {2019}
}
@techreport{Boehm2000,
author = {Boehm, Barry and Hansen, Wilfred J},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Boehm, Hansen - 2000 - Spiral Development Experience, Principles, and Refinements Spiral Development Workshop February 9, 2000.pdf:pdf},
keywords = {anchor point milestones,cyclic phases,invariant,iterative process,project management,risk management,software development,spiral development model,variant},
title = {{Spiral Development: Experience, Principles, and Refinements Spiral Development Workshop February 9, 2000}},
year = {2000}
}
@techreport{Pecka,
abstract = {We overview different approaches to safety in (semi)auto-nomous robotics. Particularly, we focus on how to achieve safe behavior of a robot if it is requested to perform exploration of unknown states. Presented methods are studied from the viewpoint of reinforcement learning, a partially-supervised machine learning method. To collect training data for this algorithm, the robot is required to freely explore the state space-which can lead to possibly dangerous situations. The role of safe exploration is to provide a framework allowing exploration while preserving safety. The examined methods range from simple algorithms to sophisticated methods based on previous experience or state prediction. Our overview also addresses the issues of how to define safety in the real-world applications (apparently absolute safety is unachievable in the continuous and random real world). In the conclusion we also suggest several ways that are worth researching more thoroughly.},
author = {Pecka, Martin and Svoboda, Tomas},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Pecka, Svoboda - Unknown - LNCS 8906 - Safe Exploration Techniques for Reinforcement Learning – An Overview.pdf:pdf},
keywords = {Safe exploration,policy search,reinforcement learning},
title = {{LNCS 8906 - Safe Exploration Techniques for Reinforcement Learning – An Overview}},
url = {http://cmp.felk.cvut.cz/∼peckama2}
}
@article{Bradley1997,
abstract = {In this paper we investigate the use of the area under the receiver operating characteristic (ROC) curve (AUC) as a performance measure for machine learning algorithms. As a case study we evaluate six machine learning algorithms (C4.5, Multiscale Classifier, Perceptron, Multi-layer Perceptron, k-Nearest Neighbours, and a Quadratic Discriminant Function) on six "real world" medical diagnostics data sets. We compare and discuss the use of AUC to the more conventional overall accuracy and find that AUC exhibits a number of desirable properties when compared to overall accuracy: increased sensitivity in Analysis of Variance (ANOVA) tests; a standard error that decreased as both AUC and the number of test samples increased; decision threshold independent; and it is invafiant to a priori class probabilities. The paper concludes with the recommendation that AUC be used in preference to overall accuracy for "single number" evaluation of machine learning algorithms.},
author = {Bradley, Andrew E},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Bradley - 1997 - THE EVALUATION OF MACHINE LEARNING ALGORITHMS.pdf:pdf},
journal = {Pattern Recognit.},
number = {7},
pages = {1145--1159},
title = {{The Use of the Area Under the ROC Curve in the Evaluation of Machine Learning Algorithms}},
volume = {30},
year = {1997}
}
@misc{Kelly2019,
abstract = {Background: Artificial intelligence (AI) research in healthcare is accelerating rapidly, with potential applications being demonstrated across various domains of medicine. However, there are currently limited examples of such techniques being successfully deployed into clinical practice. This article explores the main challenges and limitations of AI in healthcare, and considers the steps required to translate these potentially transformative technologies from research to clinical practice. Main body: Key challenges for the translation of AI systems in healthcare include those intrinsic to the science of machine learning, logistical difficulties in implementation, and consideration of the barriers to adoption as well as of the necessary sociocultural or pathway changes. Robust peer-reviewed clinical evaluation as part of randomised controlled trials should be viewed as the gold standard for evidence generation, but conducting these in practice may not always be appropriate or feasible. Performance metrics should aim to capture real clinical applicability and be understandable to intended users. Regulation that balances the pace of innovation with the potential for harm, alongside thoughtful post-market surveillance, is required to ensure that patients are not exposed to dangerous interventions nor deprived of access to beneficial innovations. Mechanisms to enable direct comparisons of AI systems must be developed, including the use of independent, local and representative test sets. Developers of AI algorithms must be vigilant to potential dangers, including dataset shift, accidental fitting of confounders, unintended discriminatory bias, the challenges of generalisation to new populations, and the unintended negative consequences of new algorithms on health outcomes. Conclusion: The safe and timely translation of AI research into clinically validated and appropriately regulated systems that can benefit everyone is challenging. Robust clinical evaluation, using metrics that are intuitive to clinicians and ideally go beyond measures of technical accuracy to include quality of care and patient outcomes, is essential. Further work is required (1) to identify themes of algorithmic bias and unfairness while developing mitigations to address these, (2) to reduce brittleness and improve generalisability, and (3) to develop methods for improved interpretability of machine learning predictions. If these goals can be achieved, the benefits for patients are likely to be transformational.},
author = {Kelly, Christopher J. and Karthikesalingam, Alan and Suleyman, Mustafa and Corrado, Greg and King, Dominic},
booktitle = {BMC Med.},
doi = {10.1186/s12916-019-1426-2},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Kelly et al. - 2019 - Key challenges for delivering clinical impact with artificial intelligence(2).pdf:pdf},
issn = {17417015},
keywords = {Algorithms,Artificial intelligence,Evaluation,Machine learning,Regulation,Translation},
month = {oct},
number = {1},
pages = {195},
pmid = {31665002},
publisher = {BioMed Central Ltd.},
title = {{Key challenges for delivering clinical impact with artificial intelligence}},
url = {https://bmcmedicine.biomedcentral.com/articles/10.1186/s12916-019-1426-2},
volume = {17},
year = {2019}
}
@article{Batarseh2021,
abstract = {Artificial Intelligence (AI) algorithms are increasingly providing decision making and operational support across multiple domains. AI includes a wide (and growing) library of algorithms that could be applied for different problems. One important notion for the adoption of AI algorithms into operational decision processes is the concept of assurance. The literature on assurance, unfortunately, conceals its outcomes within a tangled landscape of conflicting approaches, driven by contradicting motivations, assumptions, and intuitions. Accordingly, albeit a rising and novel area, this manuscript provides a systematic review of research works that are relevant to AI assurance, between years 1985 and 2021, and aims to provide a structured alternative to the landscape. A new AI assurance definition is adopted and presented, and assurance methods are contrasted and tabulated. Additionally, a ten-metric scoring system is developed and introduced to evaluate and compare existing methods. Lastly, in this manuscript, we provide foundational insights, discussions, future directions, a roadmap, and applicable recommendations for the development and deployment of AI assurance.},
author = {Batarseh, Feras A. and Freeman, Laura and Huang, Chih Hao},
doi = {10.1186/s40537-021-00445-7},
file = {:C$\backslash$:/Users/Suchendra/AppData/Local/Mendeley Ltd./Mendeley Desktop/Downloaded/Batarseh, Freeman, Huang - 2021 - A survey on artificial intelligence assurance.pdf:pdf},
issn = {21961115},
journal = {J. Big Data},
keywords = {AI assurance,Data Engineering,Explainable AI (XAI),Validation and verification},
month = {dec},
number = {1},
pages = {1--30},
publisher = {Springer Science and Business Media Deutschland GmbH},
title = {{A survey on artificial intelligence assurance}},
url = {https://doi.org/10.1186/s40537-021-00445-7},
volume = {8},
year = {2021}
}
@techreport{Serban2020,
abstract = {Deep reinforcement learning has recently shown many impressive successes. However, one major obstacle towards applying such methods to real-world problems is their lack of data-efficiency. To this end, we propose the Bottleneck Simulator: a model-based reinforcement learning method which combines a learned, factorized transition model of the environment with rollout simulations to learn an effective policy from few examples. The learned transition model employs an abstract, discrete (bottleneck) state, which increases sample efficiency by reducing the number of model parameters and by exploiting structural properties of the environment. We provide a mathematical analysis of the Bottleneck Simulator in terms of fixed points of the learned policy, which reveals how performance is affected by four distinct sources of error: an error related to the abstract space structure, an error related to the transition model estimation variance, an error related to the transition model estimation bias, and an error related to the transition model class bias. Finally, we evaluate the Bottleneck Simulator on two natural language processing tasks: a text adventure game and a real-world, complex dialogue response selection task. On both tasks, the Bottleneck Simulator yields excellent performance beating competing approaches.},
author = {Serban, Iulian Vlad and Sankar, Chinnadhurai and Pieper, Michael and Montreal, Polytechnique Montreal and Pineau, Canada Joelle and Bengio, Yoshua},
booktitle = {J. Artif. Intell. Res.},