References
- Aamodt, A., & Plaza, E. (1994). Case-based reasoning: Foundational issues, methodological variations, and system approaches. AI Communications, 7(1), 39–59.10.3233/AIC-1994-7104
- Alaieri, F., & Vellino, A. (2016). Ethical decision making in robots: Autonomy, trust and responsibility. In A. Agah, J. J. Cabibihan, A. M. Howard, M. A. Salichs, & H. He (Eds.), Social Robotics (Vol. 9979, pp. 159–168). https://doi.org/10.1007/978-3-319-47437-3_1610.1007/978-3-319-47437-3_16
- Albrecht, S. V., & Stone, P. (2018). Autonomous agents modelling other agents: A comprehensive survey and open problems. Artificial Intelligence, 258, 66–95. https://doi.org/10.1016/j.artint.2018.01.00210.1016/j.artint.2018.01.002
- Allen, C., Wallach, W., & Smit, I. (2006). Why Machine Ethics? IEEE Intelligent Systems, 21(4), 12–17. https://doi.org/10.1109/MIS.2006.8310.1109/MIS.2006.83
- Anderson, M., & Anderson, S. L. (2008). ETHEL: Toward a Principled Ethical EldercareRobot. In M. Skubic (Ed.), AI in Eldercare: New Solutions to Old Problems: Papers from the AAAI Fall Symposium. Menlo Park, California: AAAI Press.
- Anderson, M., & Anderson, S. L. (2011). Machine Ethics. Cambridge: Cambridge University Press.10.1017/CBO9780511978036
- Anderson, M., & Anderson, S. L. (2014). GenEth: General Ethical Dilemma Analyzer. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence (pp. 253–261). Retrieved January 10, 2020 from: https://www.researchgate.net/publication/261871702_GenEth_A_General_Ethical_Dilemma_Analyzer10.1609/aaai.v28i1.8737
- Arkin, R. C., & Ulam, P. (2009). An Ethical Adaptor: Behavioral Modification Derived from Moral Emotions. Retrieved January 10, 2020 from: http://www.cc.gatech.edu/ai/robot-lab/online-publications/ArkinUlamTechReport2009.pdf10.1109/CIRA.2009.5423177
- Bard, G. V. (2007). Spelling-error tolerant, order-independent pass-phrases via the damerau-levenshtein string-edit distance metric. Proceedings of the Fifth Australasian Symposium on ACSW Frontiers – Volume 68, 117–124. Retrieved January 10, 2020 from: http://dl.acm.org/citation.cfm?id=1274531.1274545
- Broadie, S. (1991). Ethics with Aristotle. New York/Oxford: Oxford University Press.
- Bostrom, N. 2003. Cognitive, Emotive and Ethical Aspects of Decision Making in Humans and in Artificial Intelligence, Vol. 2, ed. I. Smith et al., International Institute of Advanced Studies in Systems Research and Cybernetics (pp. 12–17) [revised version available at https://nickbostrom.com/ethics/ai.html].
- Cave, S., Nyrup, R., Vold, K., & Weller, A. (2019). Motivations and Risks of Machine Ethics. Proceedings of the IEEE, 107(3), 562–574. https://doi.org/10.1109/JPROC.2018.286599610.1109/JPROC.2018.2865996
- Cole, D. (2019). The Chinese Room Argument. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy (Spring 2019). Retrieved January 10, 2020 from: https://plato.stanford.edu/archives/spr2019/entries/chinese-room/
- Devlin, K. (1991). Logic and Information. New York, NY, US: Cambridge University Press.
- European Parliament. (2017, February 16). Civil Law Rules on Robotics European Parliament resolution of 16 February 2017 with recommendations to the Commission on Civil Law Rules on Robotics (2015/2103(INL)). Retrieved January 10, 2020 from: http://www.europarl.europa.eu/doceo/document/TA-8-2017-0051_EN.pdf
- Floridi, L., Sanders, J. W. (2004). On the morality of artificial agents. Minds and Machines, 14(3), 349–379. https://doi.org/10.1023/B:MIND.0000035461.63578.9d10.1023/B:MIND.0000035461.63578.9d
- Flyvbjerg, B. (2011). Making Social Science Matter: Why Social Inquiry Fails and How It Can Succeed Again (S. Sampson, Trans.). Cambridge: Cambridge University Press.
- Gilleland, M. (2006). Levenshtein Distance, in Three Flavors. Retrieved December 3, 2019 from https://people.cs.pitt.edu/~kirk/cs1501/Pruhs/Spring2006/assignments/editdistance/Levenshtein%20Distance.htm
- Graham, G. (2017). Behaviorism. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy (Spring 2017). Retrieved January 10, 2020 from: https://plato.stanford.edu/archives/spr2017/entries/behaviorism/
- Hasna, O. L., & Potolea, R. (2016). The longest common subsequence distance using a complexity factor. Proceedings of the 8th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management, 336–343. https://doi.org/10.5220/000606760336034310.5220/0006067603360343
- Hempel, C. G. (1966). Philosophy of Natural Science. Upper Saddle River, NJ: Prentice-Hall.
- Hibbard, B. (2015). Ethical Artificial Intelligence. ArXiv:1411.1373 [Cs]. Retrieved January 10, 2020 from: http://arxiv.org/abs/1411.1373.
- Jaro, M. A. (1989). Advances in record-linkage methodology as applied to matching the 1985 census of Tampa, Florida. Journal of the American Statistical Association, 84(406), 414–420. https://doi.org/10.1080/01621459.1989.1047878510.1080/01621459.1989.10478785
- Leite, I., Martinho, C., & Paiva, A. (2013). Social Robots for Long-Term Interaction: A Survey. International Journal of Social Robotics, 5(2), 291–308. https://doi.org/10.1007/s12369-013-0178-y10.1007/s12369-013-0178-y
- Navarro, G. (2001). A guided tour to approximate string matching. ACM Computing Surveys, 33(1), 31–88. https://doi.org/10.1145/375360.37536510.1145/375360.375365
- Nelson, P. (2016). One autonomous car will use 4,000 GB of data per day. Network World. Retrieved January 10, 2020 from: https://www.networkworld.com/article/3147892/one-autonomous-car-will-use-4000-gb-of-dataday.html
- Oers, R. van, & Wesselman, E. (2016). Social Robots. Retrieved January 10, 2020 from: https://assets.kpmg/content/dam/kpmg/pdf/2016/06/social-robots.pdf
- Polak, P., & Krzanowski, R. (2020). Ethics in autonomous robots as philosophy in silico: The study case of phronetic machine ethics. Logos i Ethos, 54, 33–48.10.15633/lie.3576
- Poole, D. L., & Mackworth, A. K. (2017). Artificial Intelligence: Foundations of Computational Agents (2nd edition). Retrieved January 10, 2020 from: https://artint.info/2e/html/ArtInt2e.html
- Prestes, E., Carbonera, J. L., Rama Fiorini, S., M. Jorge, V. A., Abel, M., Madhavan, R., ... & Schlenoff, C. (2013). Towards a core ontology for robotics and automation. Robotics and Autonomous Systems, 61(11), 1193–1204. https://doi.org/10.1016/j.robot.2013.04.00510.1016/j.robot.2013.04.005
- Randall, J. H. (1965). Aristotle (4th ed.). New York: Columbia University Press.
- Reeve, C. D. C. (1992). Practices of Reason: Aristotle’s Nicomachean Ethics. Oxford: Clarendon Press.
- Richter, M. M., & Weber, R. (2013). Case-Based Reasoning: A Textbook. https://doi.org/10.1007/978-3-642-40167-110.1007/978-3-642-40167-1
- Searle, J. R. (1998). Mind, Language and Society: Philosophy in the Real World. New York: Basic Books.
- Tarleton, N. (2010). Coherent Extrapolated Volition: A Meta-Level Approach to Machine Ethics. San Francisco, CA: The Singularity Institute.
- Tenorth, M., & Beetz, M. (2017). Representations for robot knowledge in the KnowRob framework. Artificial Intelligence, 247, 151–169. https://doi.org/10.1016/j.artint.2015.05.01010.1016/j.artint.2015.05.010
- Väyrynen, P. (2017). Thick Ethical Concepts. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy (Fall 2017). Retrieved from https://plato.stanford.edu/archives/fall2017/entries/thick-ethical-concepts/
- Wallach, W., & Allen, C. (2009). Moral Machines: Teaching Robots Right from Wrong. Oxford: Oxford University Press.10.1093/acprof:oso/9780195374049.001.0001
- Wallach, W., Allen, C., & Smit, I. (2008). Machine morality: Bottom-up and top-down approaches for modelling human moral faculties. AI & SOCIETY, 22(4), 565–582. https://doi.org/10.1007/s00146-007-0099-010.1007/s00146-007-0099-0
- Wallach, W., Franklin, S., & Allen, C. (2010). A conceptual and computational model of moral decision making in human and artificial agents. Topics in Cognitive Science, 2(3), 454–485. https://doi.org/10.1111/j.1756-8765.2010.01095.x10.1111/j.1756-8765.2010.01095.x
- Winkler, W. E. (1990). String Comparator Metrics and Enhanced Decision Rules in the Fellegi-Sunter Model of Record Linkage. 354–359.
- Yudkowsky, E. (2004). Coherent Extrapolated Volition. San Francisco, CA: The Singularity Institute