Ai4lam Look Book
A Growing Knowledge Base of AI �Projects in Libraries, Archives and Museums
GallicaSNOOP (2018-2020)
R&D project based on the visual similarity search engine “SNOOP”, developed by INRIA (FR national computer science lab) and INA research team (FR audiovisual heritage agency)
GallicaSNOOP experiments
on a 1M Gallica images dataset scraped with IIIF
SNOOP
SNOOP is the Pl@ntNet app search engine
�
https://plantnet.org/
GallicaSNOOP proof of concept
Query=press �agency photo (Gallica,1912)
Results=reproductions �in newspapers (1912)
“Human in the loop” query: iterating on results
First query=user �photos (2020)
Results=heritage photos �(Gallica, 1910-1920)
Transkribus�A platform for the transcription, recognition and searching of historical documents
Günter Mühlberger
Digitisation and Digital Preservation group
University of Innsbruck
Dirc Jansz �Schiouwer � Op hHuijden den 18en. Decemb. @ 1638. � Compareerde voor mij Hendrick Schaef � Notaris pPub. etc. Trijn Barents huijsvr: van
Rules of thumb
Google's Cloud Offering
Generic OCR API for 232 languages across 31 scripts
HTR support for languages written in Latin, Japanese, and Korean scripts
Unified model recognizing both handwriting and block printed
Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem
Proprietary + Confidential
Google's Cloud Offering
Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem
Proprietary + Confidential
OCR for Bangla – The Challenge
12
bl.uk/early-indian-printed-books @BL_IndianPrint tom.derrick@bl.uk
Typical page from BL Bengali Books
www.bl.uk
13
Initiatives to find Bangla OCR solution
ICDAR Competitions 2017 & 2019
Transkribus
bl.uk/early-indian-printed-books | primaresearch.org/datasets/REID2019 | doi.org/10.23536/505 | tom.derrick@bl.uk
primaresearch.org/datasets/REID2019 | ICDAR dataset: doi.org/10.23536/505 | Transkribus dataset: doi.org/10.23636/506
www.bl.uk
14
Wikisource Transcriptions
https://commons.wikimedia.org/wiki/Category:Two_Centuries_of_Indian_Print
https://commons.wikimedia.org/wiki/Category:Two_Centuries_of_Indian_Print
www.bl.uk
http://www.robots.ox.ac.uk/~vgg/
Role of the VGG Digital Humanities Ambassadorship
“To disseminate VGG research to appropriate communities…” and to feed back research questions, interesting datasets, ideas for new development, bug reports, feature requests etc…
�
VGG implementations: searching by instance (left) and category (right)
Other VGG demos - http://www.robots.ox.ac.uk/~vgg/demo
SARAH system (2018-ongoing)
Computer vision for automated tag creation in archives
SARAH system (2018-ongoing) - overview
SARAH system (2018-ongoing) - historical figures identification
Common Crawl News
Common Crawl Stories
Open WebText
The Colossal Norwegian Corpus
Lessons Learned
https://github.com/NBAiLab/notram/
AI-Lab (from North to South)
Norwegian Transformer Model
&
Colossal Norwegian Corpus
per.kummervold@nb.no
Project State (15.12.20)
pit.schneider@bnl.etat.lu
pit.schneider@bnl.etat.lu
The Netherlands Institute for Sound and Vision is a use case partner on AI for Social Sciences and Humanities research into issues of bias, framing and representation in media.��Resulting in user requirements research, validation and demonstration of AI tooling that is Trustworthy, Interoperable and enables multimodal media analysis in a configurable manner.
26
2020 - 2024
Philo van Kemenade
pvkemenade@beeldengeluid.nl
Distributed Annotation ‘n’ Enrichment (DANE)
The Distributed Annotation ‘n’ Enrichment (DANE) system handles compute task assignment and file storage for the automatic annotation of content.
The use-case for which DANE was designed centres around the issue that the compute resources, and the collection of source media are not on the same device. Due to limited resources or policy choices it might not be possible or desirable to bulk transfer all source media to the compute resources, alternatively the source collection might be continuously growing or require on-demand processing.
��Development by Nanne van Noord�https://dane.readthedocs.io/en/latest/index.html
Text detection
Object detection
Classification
Indexing process automation
Project description
In 2014, the French National Audiovisual start to redesign its whole information system, centralizing databases and harmonizing data models in order to provide and maintain data consistency.
in 2019, the architecture of this whole new information system was completed. This allowed us to work on AI-based solutions in order to fully or partially automate segmentation and indexing process of TV programs.
Institut National de l’Audiovisuel - 94360 Bry-Sur-Marne, France
A toolbox made for set up AI workflows
What have we learned?
Some takeaways from the work :
Always think usecase first
Build transversal teams
Involve final users
Build human-computer interfaces
Combine tools and stay flexible
Centralize systems and models
Keep data consistency
Segmentation and indexing of daily broadcast on news channels
ReTV: Bringing Archival Content to Audiences Online
https://www.visualcapitalist.com/media-consumption-covid-19/
Rasa Bocyte rbocyte@beeldengeluid.nl @rasa_bocyte
32
👉 Online Demo for Video Summarisation: http://multimedia2.iti.gr/videosummarization/service/start.html 👈
Content Wizard Prototype
Rasa Bocyte rbocyte@beeldengeluid.nl @rasa_bocyte
AV Archives in Your Pocket | 4u2 Messenger Prototype
Rasa Bocyte rbocyte@beeldengeluid.nl @rasa_bocyte
AMP: Audiovisual Metadata Platform
Challenge: Abundance of digitized and born-digital AV media
Proposed solution: Leverage automation / machine learning together with human expertise to produce more efficient workflows
Media Content
Existing Metadata
Workflow system
MGM
MGM
MGM
Enriched Metadata
Target System
Users
AMP
Current Phase: AMP Pilot Development (AMPPD)
More information at https://go.iu.edu/amppd�Twitter: @AVMetadata �Contact: Jon Dunn, jwd@iu.edu
Current Phase: AMP Pilot Development (AMPPD)
More information at https://go.iu.edu/amppd�Twitter: @AVMetadata �Contact: Jon Dunn, jwd@iu.edu
Current Phase: AMP Pilot Development (AMPPD)
More information at https://go.iu.edu/amppd�Twitter: @AVMetadata �Contact: Jon Dunn, jwd@iu.edu
Dr. Oonagh Murphy, Goldsmiths, University of London
Dr. Elena Villaespesa, Pratt Institute
The Museums and Artificial Intelligence Network brought together a range of senior museum professionals and prominent academics to develop the conversation around AI, ethics and museums. This project was funded by the AHRC.
Through a series of industry workshops in London, New York and San Diego, the network facilitated in depth discussions designed to open up debate around the key parameters, methods and paradigms of AI in a museum context.
4378503678
AI + Visitor data
AI + Collection data
Do museums have the necessary data governance and processes in place to manage AI?
How does the current museum sector code of ethics and regulations cover the rapid growing AI field?
What are the best ethical practices to collect and analyze data with AI?
What skills might museum workers need to have to work with AI to get visitor insights?
What are the opportunities and challenges to apply AI technologies to collections data?
How can museums minimize algorithm biases to interpret their collections?
Would the lack of diversity in the museum and AI fields be reflected in the outcomes of using these technologies?
What are the implications of museums engaging with big tech companies?
4378503678
Toolkit
Discovering Pathways Through Collections:
A Museum Recommender System
Recommender system using digital collections:
Museum Recommender Engine for Collections
Why might all this ML be useful?
Lukas Noehrer [@LukasNoehrer] [lukas.noehrer@manchester.ac.uk]
Discovering Pathways Through Collections:
A Museum Recommender System
Data collection and entry define outcome of computational process
Reverse collecting
Questioning the dataset is pivotal to critically address:
- “history of elitism and exclusion”
- bias, racism, and inequalities in collections
ML community needs fair and ethical approaches: avoid ‘blind ingestions of data’ -> downstream harm
ML can help to explore and create new knowledge
Caveat of fallacies: true interpretation? Algorithms are just as biased as the data used
Reproducibility and openness are key
Most commercial systems trim data to suit certain audiences
Define according to data available rather than to trim the data: avoid data cosmetics
Consider different needs of audiences and their information seeking behaviour
Outside the search box exploration beyond authoritative narratives and fictitious ‘neutral’ displays
Develop together with audiences from the beginning
Data can hardly ever be used in its original state
Feature engineering presents challenges and complexities
Interdisciplinary effort
Curatorial trap
Documentation of what was added/manipulated/deleted
Choice of model and algorithm needs careful consideration and reflection
Power and limitations of ML techniques
Aim:
(i) relevant objects,
(ii) novel data,
(iii) serendipitous effects, and
(iv) diverse range
Not all models feasible due to data requirements, scalability, and affordability
System can be evaluated with statistical methods, i.e. accuracy and model performance
User experience might be a better indicator for museum applications
Trade-offs (model vs. user-centric)
User-centric methods
Content-based: TF-IDF and Conv. Autoencoder
Discovering Pathways Through Collections:
A Museum Recommender System
MAKING SMITHSONIAN OPEN ACCESS ACCESSIBLE WITH PYTHON AND DASK
SI OPEN ACCESS RELEASE
Of the Smithsonian’s 155 million objects, 2.1 million library volumes and 156,000 cubic feet of archival collections:
3 WAYS TO ACCESS