|
2 | 2 | layout: reference |
3 | 3 | --- |
4 | 4 |
|
5 | | -## Course Learner Profiles |
| 5 | +## Course Learner Profiles |
| 6 | + |
| 7 | +### Kunal the Curator |
6 | 8 |
|
7 | | -###Kunal the Curator |
8 | 9 | Kunal is a newly hired curator of South Asian Collections at a national library. He holds a PhD in Bengali Language and Literature and has responsibility for cataloguing and acquisitions. His main tasks right now are to get a good understanding of the extent and range of the vast historical collection he is looking after, no easy task as the collection has been acquired over many decades, is potentially in the tens of thousands of items, and catalogue records are of varied quality. He knows that there are digital skills and tools that could help him more systematically approach this work of identifying gaps in the collection as well as improving existing catalogue records. |
9 | 10 |
|
10 | 11 | He is comfortable using tools like Excel and has recently taught himself how to use OpenRefine in order to clean up and try to analyse tens of thousands of bibliographic records he’s exported from the catalogue in .xls. Though he’s found OpenRefine useful for normalising batches of records and some simple analysis, he feels he has reached its limits. He is interested in going further and looking at machine learning approaches such as Natural Language Processing for better categorising the digitised Bengali printed books, for instance, which he’s found have very minimal descriptions and either poor or no OCR at all. |
|
31 | 32 | : |
32 | 33 |
|
33 | 34 | machine learning |
34 | | -: The study or use of algorithms whose performance improves as they are given more data. Machine learning algorithms often use training data to build a model. Their performance is then measured by how well they predict the properties of test data. It is a set of technologies and methods for finding rules when they are too complex to define. They are systems which find rules, learn, and make predictions from data without being explicitely programmed to do so. https://glosario.carpentries.org/en/#machine_learning |
| 35 | +: The study or use of algorithms whose performance improves as they are given more data. Machine learning algorithms often use training data to build a model. Their performance is then measured by how well they predict the properties of test data. It is a set of technologies and methods for finding rules when they are too complex to define. They are systems which find rules, learn, and make predictions from data without being explicitely programmed to do so. <https://glosario.carpentries.org/en/#machine_learning> |
35 | 36 |
|
36 | 37 | model |
37 | | -: A specification of the mathematical relationship between different variables.https://glosario.carpentries.org/en/#model |
| 38 | +: A specification of the mathematical relationship between different variables. <https://glosario.carpentries.org/en/#model> |
38 | 39 |
|
39 | 40 | regression analysis : |
40 | 41 |
|
41 | 42 | reinforcement learning |
42 | | -: Any machine learning algorithm which is not given specific goals to meet, but instead is given feedback on whether or not it is making progress. https://glosario.carpentries.org/en/#reinforcement_learning |
| 43 | +: Any machine learning algorithm which is not given specific goals to meet, but instead is given feedback on whether or not it is making progress. <https://glosario.carpentries.org/en/#reinforcement_learning> |
43 | 44 |
|
44 | 45 | semi-supervised learning : |
45 | 46 |
|
46 | 47 | supervised learning |
47 | | -: A machine learning algorithm in which a system is taught to classify values given training data containing previously-classified values. https://glosario.carpentries.org/en/#supervised_learning |
| 48 | +: A machine learning algorithm in which a system is taught to classify values given training data containing previously-classified values. <https://glosario.carpentries.org/en/#supervised_learning> |
48 | 49 |
|
49 | 50 | test data |
50 | | -: Test data is a portion of a dataset used to evaluate the correctness of a machine learning algorithm after it has been trained. It should always be separated from the training data to ensure that the model is properly tested with unseen data. https://glosario.carpentries.org/en/#test_data |
| 51 | +: Test data is a portion of a dataset used to evaluate the correctness of a machine learning algorithm after it has been trained. It should always be separated from the training data to ensure that the model is properly tested with unseen data. <https://glosario.carpentries.org/en/#test_data> |
51 | 52 |
|
52 | 53 | training data |
53 | | -: Training data is a portion of a dataset used to train machine learning algorithm to recognise similar data. It should always be separated from the test data to ensure that the model is properly tested with data it has never seen before. https://glosario.carpentries.org/en/#training_data |
| 54 | +: Training data is a portion of a dataset used to train machine learning algorithm to recognise similar data. It should always be separated from the test data to ensure that the model is properly tested with data it has never seen before. <https://glosario.carpentries.org/en/#training_data> |
54 | 55 |
|
55 | 56 | unsupervised learning |
56 | | -: Algorithms that cluster data without knowing in advance what the groups will be. https://glosario.carpentries.org/en/#unsupervised_learning |
| 57 | +: Algorithms that cluster data without knowing in advance what the groups will be. <https://glosario.carpentries.org/en/#unsupervised_learning> |
57 | 58 |
|
58 | 59 | ## External References: Resources Consulted & Recommended Reading |
59 | 60 |
|
60 | 61 | Ameisen, Emmanuel. Building Machine Learning Powered Applications: Going from Idea to Product, 2020. |
61 | 62 |
|
62 | | -Barbosa, N., & Chen, M. (2021). Rehumanized Crowdsourcing. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. Dl.acm.org. Retrieved 29 March 2021, from https://dl.acm.org/doi/10.1145/3290605.3300773. |
| 63 | +Barbosa, N., & Chen, M. (2021). Rehumanized Crowdsourcing. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. Dl.acm.org. Retrieved 29 March 2021, from <https://dl.acm.org/doi/10.1145/3290605.3300773>. |
63 | 64 |
|
64 | | -Barlow, R. (2014). BU Research: A Riddle Reveals Depth of Gender Bias. BU Today. Boston University. Retrieved 29 March 2021, from https://www.bu.edu/articles/2014/bu-research-riddle-reveals-the-depth-of-gender-bias. |
| 65 | +Barlow, R. (2014). BU Research: A Riddle Reveals Depth of Gender Bias. BU Today. Boston University. Retrieved 29 March 2021, from <https://www.bu.edu/articles/2014/bu-research-riddle-reveals-the-depth-of-gender-bias>. |
65 | 66 |
|
66 | | -Catanzaro, B. (2019, December 4). “Datasets make algorithms: how creating, curating, and distributing data creates modern AI.” [Video file]. Retrieved from https://library.stanford.edu/projects/fantastic-futures. |
| 67 | +Catanzaro, B. (2019, December 4). “Datasets make algorithms: how creating, curating, and distributing data creates modern AI.” [Video file]. Retrieved from <https://library.stanford.edu/projects/fantastic-futures>. |
67 | 68 |
|
68 | | -Coleman, C. (2020). Managing Bias When Library Collections Become Data. International Journal Of Librarianship, 5(1), 8-19. https://doi.org/10.23974/ijol.2020.vol5.1.162. |
| 69 | +Coleman, C. (2020). Managing Bias When Library Collections Become Data. International Journal Of Librarianship, 5(1), 8-19. <https://doi.org/10.23974/ijol.2020.vol5.1.162>. |
69 | 70 |
|
70 | | -Cordell, Ryan. ‘Machine Learning + Libraries’. LC Labs. Accessed 28 March 2021. https://labs.loc.gov/static/labs/work/reports/Cordell-LOC-ML-report.pdf. |
| 71 | +Cordell, Ryan. ‘Machine Learning + Libraries’. LC Labs. Accessed 28 March 2021. <https://labs.loc.gov/static/labs/work/reports/Cordell-LOC-ML-report.pdf>. |
71 | 72 |
|
72 | | -Ekowo, M. (2016). Why Numbers can be Neutral but Data Can’t. New America. Retrieved 29 March 2021, from https://www.newamerica.org/education-policy/edcentral/numbers-can-neutral-data-cant/. |
| 73 | +Ekowo, M. (2016). Why Numbers can be Neutral but Data Can’t. New America. Retrieved 29 March 2021, from <https://www.newamerica.org/education-policy/edcentral/numbers-can-neutral-data-cant/>. |
73 | 74 |
|
74 | | -Gebru, T., Morgenstern, J., Vecchione, B., Vaughan, J., Wallach, H., Daumeé III, H., & Crawford, K. (2020). Datasheets for Datasets. arXiv.org. Retrieved 29 March 2021, from https://arxiv.org/abs/1803.09010v3. |
| 75 | +Gebru, T., Morgenstern, J., Vecchione, B., Vaughan, J., Wallach, H., Daumeé III, H., & Crawford, K. (2020). Datasheets for Datasets. arXiv.org. Retrieved 29 March 2021, from <https://arxiv.org/abs/1803.09010v3>. |
75 | 76 |
|
76 | | -Hellström, T., Dignum, V., & Bensch, S. (2020). Bias in Machine Learning What is it Good (and Bad) for?. arXiv preprint. Retrieved 20 April 2021, from https://arxiv.org/abs/2004.00686v2. |
| 77 | +Hellström, T., Dignum, V., & Bensch, S. (2020). Bias in Machine Learning What is it Good (and Bad) for?. arXiv preprint. Retrieved 20 April 2021, from <https://arxiv.org/abs/2004.00686v2>. |
77 | 78 |
|
78 | 79 | Howard, Jeremy, Sylvain Gugger, and an O’Reilly Media Company Safari. Deep Learning for Coders with Fastai and PyTorch, 2020. |
79 | 80 |
|
80 | | -Jo, E., & Gebru, T. (2020). Lessons from archives. Proceedings Of The 2020 Conference On Fairness, Accountability, And Transparency. https://doi.org/10.1145/3351095.3372829. |
| 81 | +Jo, E., & Gebru, T. (2020). Lessons from archives. Proceedings Of The 2020 Conference On Fairness, Accountability, And Transparency. <https://doi.org/10.1145/3351095.3372829>. |
81 | 82 |
|
82 | 83 | Lakshmanan, Valliappa, Sara Robinson, Michael Munn, and an O’Reilly Media Company Safari. Machine Learning Design Patterns, 2021. |
83 | 84 |
|
84 | | -Mayson, Sandra Gabriel, Bias In, Bias Out (2019). 128 Yale Law Journal 2218, University of Georgia School of Law Legal Studies Research Paper No. 2018-35, Available at SSRN: https://ssrn.com/abstract=3257004. |
| 85 | +Mayson, Sandra Gabriel, Bias In, Bias Out (2019). 128 Yale Law Journal 2218, University of Georgia School of Law Legal Studies Research Paper No. 2018-35, Available at SSRN: <https://ssrn.com/abstract=3257004>. |
85 | 86 |
|
86 | | -Mitchell, Margaret, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, Inioluwa Deborah Raji, and Timnit Gebru. ‘Model Cards for Model Reporting’. Proceedings of the Conference on Fairness, Accountability, and Transparency, 29 January 2019, 220–29. https://doi.org/10.1145/3287560.3287596. |
| 87 | +Mitchell, Margaret, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, Inioluwa Deborah Raji, and Timnit Gebru. ‘Model Cards for Model Reporting’. Proceedings of the Conference on Fairness, Accountability, and Transparency, 29 January 2019, 220–29. <https://doi.org/10.1145/3287560.3287596>. |
87 | 88 |
|
88 | | -Omoju Miller. ‘The Myth of Innate Ability in Tech’. Accessed 20 March 2021. http://omojumiller.com/articles/The-Myth-Of-Innate-Ability-In-Tech. |
| 89 | +Omoju Miller. ‘The Myth of Innate Ability in Tech’. Accessed 20 March 2021. <http://omojumiller.com/articles/The-Myth-Of-Innate-Ability-In-Tech>. |
89 | 90 |
|
90 | | -Padilla, T. (2019). Responsible Operations: Data Science, Machine Learning, and AI in Libraries. OCLC Research Position Paper. https://doi.org/10.25333/xk7z-9g97. |
| 91 | +Padilla, T. (2019). Responsible Operations: Data Science, Machine Learning, and AI in Libraries. OCLC Research Position Paper. <https://doi.org/10.25333/xk7z-9g97>. |
91 | 92 |
|
92 | | -Slee, Tom. ‘The Incompatible Incentives of Private Sector AI’. Tom Slee, 31 March 2019. https://tomslee.github.io/publication/oup_private_sector_ai/. |
| 93 | +Slee, Tom. ‘The Incompatible Incentives of Private Sector AI’. Tom Slee, 31 March 2019. <https://tomslee.github.io/publication/oup_private_sector_ai/>. |
93 | 94 |
|
94 | | -Suresh, Harini, and John V. Guttag. ‘A Framework for Understanding Unintended Consequences of Machine Learning’. ArXiv:1901.10002 [Cs, Stat], 17 February 2020. http://arxiv.org/abs/1901.10002. |
| 95 | +Suresh, Harini, and John V. Guttag. ‘A Framework for Understanding Unintended Consequences of Machine Learning’. ArXiv:1901.10002 [Cs, Stat], 17 February 2020. <http://arxiv.org/abs/1901.10002>. |
95 | 96 |
|
96 | | -Thomas, Rachel. ‘The Problem with Metrics Is a Big Problem for AI · Fast.Ai’. fast.ai blog. Accessed 18 March 2021. https://www.fast.ai/2019/09/24/metrics/. |
| 97 | +Thomas, Rachel. ‘The Problem with Metrics Is a Big Problem for AI · Fast.Ai’. fast.ai blog. Accessed 18 March 2021. <https://www.fast.ai/2019/09/24/metrics/>. |
97 | 98 |
|
98 | 99 | {% include links.md %} |
0 commit comments