Automatic Title Translations of Artworks

INSIGHT aims to use recent advances in Artificial Intelligence (language technology and computer vision) to support the enrichment of art collections with descriptive metadata. Multilingualism is a crucial research aspect in this respect. For the Royal Museums of Fine Arts of Belgium it is very important to be able to offer the entirely bilingual collection database (DU – FR)  in a trilingual way (DU – FR – EN). The INSIGHT project responds to this by focusing on automatic title translation of works of art.

Tower of Babel, Joos II de Momper; figures attributed to Frans II Francken, RMFAB, inv. 8032 / photo by J. Geleyns

It is not easy to tackle title translation within the art sector, even for experienced translators. Deploying machine translation for this purpose is effectively a major challenge. In some cases, visual data can help (via multimodal learning, combining computer vision and natural language processing). In this respect, we mainly think of old masterpieces with descriptive titles for more realistic scene depiction. For modern and contemporary art, the use of visual context becomes more difficult, as the link between title and work of art is often less obvious.

[Road to Bethlehem], Joshua Neustein, (1981-1983), RMFAB, inv. 10474 / photo by J. Geleyns. English title not yet validated. Official dutch title: “Weg naar Bethlehem”.

Methodology

The Departments of Linguistics and Literature at the University of Antwerp, which deals with these issues in the project, obviously does not aim to compete with commercial translation masters who have unlimited computer power. The team wants to tackle smaller, domain-specific datasets and experiment with other translation techniques such as ‘character level translation’ and ‘multimodal learning’. A number of papers were recently published in this area, setting the stage for the application of these methods “in the wild”, i.e. for actual digital collections in the GLAM sector. Character-level machine translation seems very promising in this respect, although this is computationally challenging. Interestingly, we can directly inject domain-specific knowledge that is available about artworks, such as IconClass-codes, to improve the results .


Within our specific subproject, Dutch titles form the basis for the English automated translations. The algorithms need to be trained with sufficient data for this purpose. They were pre-trained on the large EuroParl corpus and then trained with RKD datasets and some 1000 titles from the RMFAB collections which are already have available in English. Initial tests show that, as was to be expected, a much as possible sector-specific training data is needed to achieve an optimal result.

Challenge in Consistency

In order to be able to provide additional material to help raise the learning curves of the algorithms, the Museum needs to review its internal linguistic conventions (e.g. to use gender neutral terms: De Baadsters -> Bathers; to avoid initial articles, unless it is commonly used and confusion could result if it were omitted: e.g. Het meisje -> Girl; to use Self-Portrait instead of Autoportrait…) and to apply those to the existing datasets and new data input to deliver. One cannot expect the algorithms to deliver good quality if the training data shows inconsistencies. The consistent application of internal conventions is moreover important both in the initial allocation of titles (in DU and FR) and in the translation of titles.

In addition to cleaning up the existing datasets, the regulatory framework will form the basis for the volunteers who are willing to translate large quantities of titles into English in order to provide the algorithms with new training material. It is, of course, essential and inevitable that future automated title suggestions will have to pass the all-seeing eye of the curators. They will make the final adjustments to arrive at official titles validated by the Institution. Through INSIGHT we hope to, at least, facilitate the translation process by offering a first basic translation.

By: Lies Van de Cappelle (Royal Museums of Fine Arts of Belgium, Brussels)

Insight into Minerva

Recently, the INSIGHT project was invited to present its work in the DH lunch lectures series organized by our kind colleagues in Krakow. Because of the global lockdown, we presented our work virtually via a prerecorded session (archived link), followed by a live Q&A session over videoconferencing. Those who missed the talk are welcome to watch the recording which was published on YouTube afterwards. As will become clear in the talk (also embedded below), we discussed our work on Minerva, in which we capitalize on object detection methods from computer to detect musical instruments in unrestricted artworks from the visual artworks. We have an awesome paper coming up on this subject in a special issue of the journal Digital Humanities Quarterly and we’ll make sure to blog more extensively about that once it’s out — any day now. Stay tuned!

Hunting animals in art (but not for real!)

In this blog post, project member Marie Coccriamont writes about her experiences and adventures collecting and annotating images of mammals in the visual arts.

A new challenge

After the annotation process of musical instruments (a blog post will follow soon!) we searched for a new annotation subject to further continue the INSIGHT project.

Animals in art have been a popular study topic throughout the past centuries. They are represented within a wide range of art objects and art genres and have either a central or peripheral role in the composition. Animals can function as complemental elements in landscapes, genres or religious scenes, they feature prominently as heraldic charges, they have been important subjects in sculpture, etc. These representations can have different levels of meaning and can assume a symbolic function in expressing the urges of society from a particular period in time and from a particular culture.

The previously mentioned properties show that this topic can serve as an interesting and useful research subject for the INSIGHT project. However, a feasibility study identified a number of problems. What to do with the representation of beasts and mythical animals? What to do with the variety of birds, the variety of dog breeds, the variety of insects, … A demarcation on the subject was needed. Because of the specific aim of the project, specific and clear labels were necessary. With the guidance of Imagenet and Wordnet, which provide a reliable structure, we came up with a new subject: ‘mammals’. With the help of standard ontologies such as Garnier and IconClass, a research of the most common mammals in the collections of KMSKB and KMKG, and the analysis of the symbolic value of mammals throughout art history, 25 labels were chosen. Regarding these 25 labels, images of art objects were gathered from the collections of KMSKB, KMKG, Wikimedia and Rijksmuseum.

Where to get data?

A first annotation cycle was done on the basis of the collections of the KMSKB and KMKG. With the help of the existing thesaurus of both collections, images of art objects of the general category “mammals”, were collected, without taking in account the 25 chosen labels. Therefore an overview of the representation of mammals in the collections was possible. An additional collection cycle was performed using the Wikimedia and Rijksmuseum collections. I gathered at least 100 pictures of art objects per label from both collections. Therefore I could ensure an equal number of images per label.

I started off with the annotation of the collections of the KMSKB and KMKG. I came across a massive number of horses, domestic dogs, cows and sheep. The number was that big that I had to delay the annotations of those mammals to guard the equality in relation to the other mammals (see figures). Other problems that occurred, were difficulties with distinguishing mammals. In one of the examples below, we see a painting in which several horses are pulling a carriage. Because of the angle of the painting, it is difficult to annotate each separate horse using a rectangle. In yet another example, we see a painting with the representation of a group of cows and a herd of sheep. Because of the amount of cows and sheep and the bad quality of the image, it is difficult to distinguish each separate mammal.


More difficulties…

Another interesting difficulty was the representation of “fantasy” mammals. Mammals that are abstracted using the imaginative mind. Examples of these “fantasy” mammals can be found in the example below. Because of the huge unambiguity associated with these animals, I chose not to annotate these images.


Another difficulty that I came across, was a low visibility or an ambiguity of the representation of mammals. This can be related to the bad quality of the medium, the abstraction of the mammal, the angle of the art object, … You name it!

The large amount of low-quality quality images made the annotation process more difficult:

Also, and as could be anticipated,  I also experienced difficulties with labelling certain mammals. It was only during the annotation process that I realized how certain labels were too closely related to each other. Bull and cow for example, are difficult to distinguish. Rat and mice also have much similarities:

Finally, I came across several images that used mammals “as a metaphor”. Such personifications of mammals can be seen in the following images. In these images I doubted how to annotate the mammals. I choose to annotate only the head of the mammals, because the bodies are often those of humans.

We thank Marie for her huge contribution to the project and we are looking to what we will get out this wonderful data in a next phase of the project!

Matthia at ECCV

Computer Vision for Art and the ECCV conference in Munich

Check out the original paper:  “Deep Transfer Learning for Art Classification Problems”: M. Sabatelli et al. European Conference on Computer Vision (ECCV), 4th Workshop on Computer VISion for ART Analysis (VISART IV), München (GE). Code, data and camera-ready manuscript are available from Github.

The research which has so far been done in the Walloon part of the country has investigated whether state of the art algorithms that come from the fields of machine learning and Computer Vision (CV) could also be used when dealing with images representing Digital Heritage. If on the one hand, the field of CV is a well established research domain, which over the recent years has shown the effectiveness of algorithms such as Deep Convolutional Neural Networks (DCCNs), the potential of such neural architectures in the field of Digital Heritage is still an open question. Thus we have performed an empirical study that aimed to answer this research question. We have investigated a set of different strategies that can be used for training the neural networks, which, once established, have led to the conclusion that DCNNs can be a powerful tool for automatically classifying heritage objects. Besides providing an interesting support tool for art historians who aim to classify artworks, we have also tried to provide insights about why such neural networks can perform so well if properly trained, with the goal of making such algorithms understandable to the people who are less familiar with the field of machine learning.

The results of this research have led to the publication of the first CV article within the INSIGHT project called “Deep Transfer Learning for Art Classification Problems”. The paper [1] has been presented on the 9th of September at the European Conference on Computer Vision (ECCV) in Munich, during the VisArt workshop, a bi-annual event which gathers international researchers working at the intersection between Artificial Intelligence and cultural heritage. The paper has been reserved a 20 minutes talk during the workshop’s morning session called “Deep in Art”, and has been successfully accepted by the AI community.

This German experience has seen our PhD student Matthia leaving Liege at 5 in the morning of September 7th on the way to the south of Germany by train. Once settled down in his Airbnb, it was already time for him to start attending the first day of workshops at the conference, and to scientifically investigate whether German beers are as good as they say.

After having assessed the superiority of Belgian beers over the German ones, on September 9th the paper has been presented at the workshop in front of over 150 people. Besides being accepted with a lot of enthusiasm by the people attending the workshop, the paper has also caught the attention of the Google-AI headquarters in Munich. In fact they have invited Matthia to attend a special barbecue, hosted in their offices, that has given our PhD student the chance to advertise the INSIGHT project inside the biggest tech company in the world.

Meet our new team members

In the past months, thee amazing new team members started on the project. Below, they introduce themselves and their work.

Matthia Sabatelli (University of Liège)

I am working on the INSIGHT project from the Walloon part of the country where I am currently a PhD candidate in Machine Learning at the University of Liege, under the supervision of Dr. Pierre Geurts. Before that, I received my Master of Science in Artificial Intelligence (AI) from the University of Groningen in the Netherlands. My main research interests lie in the study of Deep Artificial Neural Networks where I am trying to gain insights about what makes these complex algorithms be such a successful and powerful tool in the fields of AI and Computer Vision. This understanding is then used within the INSIGHT project to build a new generation of machine learning based tools that can enrich heritage collections. Next to that, I am fascinated by the capabilities that Deep Neural Networks have in developing the most varied creative skills, ranging from the generation of artificial art, to the evolution of creative strategies that allow them to master video and board-games. More particularly, I investigate these creative processes from a Deep Reinforcement Learning perspective where I try to understand what makes Deep Neural Networks develop such original policies.

Odile Keromnes (Royal Museums of Fine Arts of Belgium)


Odile Keromnes is a young half-French half-German photographer, who lives in Brussels. Her two main passions are photography and museums. For that reason she studied art history at the university before entering the Ecole Nationale Supérieure Louis-Lumière in Paris, where she did a Master’s Degree in Photography. In this technical school, she specialized in cultural heritage digitization, doing internships in the C2RMF laboratory at the Louvre and in the photographic service of the Centre Pompidou. Now she is able to combine her two passions, working for the Royal Museums of Fine Arts of Belgium as a technical specialist for the Photographic Service. She will contribute to INSIGHT by tackling the more technical aspects of cultural heritage digitization. Her background in art history will allow her to do quality control on the first results of the AI algorithms. She will participate in the encoding of missing data for the pilot. Her work will also include the digitization of more works of art, which will enrich the project with new pictures, giving the AI more material to analyze.

Nikolay Banari (University of Antwerp)


I studied applied mathematics and physics at Moscow Institute of Physics and Technology and artificial intelligence at KU Leuven. When I was a student, I taught mathematics and physics on a part-time basis for children seeking admission to universities. Also, I worked as a Data Analyst at Forum Investment Company. Nowadays, I am a PhD researcher at University of Antwerp. I have research experience in image processing. My general scientific interests lie with the intersection of machine learning and humanities. I find the idea of applying machine learning algorithms in humanities very promising.

AI and the linking of digital heritage data – 9 November 2017, Brussels

  • The recently started BELSPO-funded INSIGHT project (Intelligent Neural Systems as Integrated Heritage Tools) organizes a launch event on 9 November 2017. This event will take the form of an afternoon of plenaries by internationally recognized speakers on topics relating to Artificial Intelligence, Heritage data and Digital Art history. This afternoon will take place at the Musical Instruments Museum in Brussels (Hofbergstraat 2, Brussels). Afterwards you are cordially invited to a reception. Registration is free but participants are invited to register through sending an email to mike.kestemont@uantwerp.be. The presentation of the event will be in the hands of Bart Magnus (PACKED).

Schedule

Welcome (Mike Kestemont) [slides]

13:00-13:45 Seth van Hooland (Université Libre de Bruxelles): Understanding the perils of Linked Data through the history of data modeling
13:45-14:30 Benoit Seguin (École Polytechnique fédérale de Lausanne): The Replica Project: Navigating Iconographic Collections at Scale
14:30-15:15 Roxanne Wyns (KULeuven): International Image Interoperability Framework (IIIF). Sharing high-resolution images across institutional boundaries
15:15-15:45 Break
15:45-16:30 Saskia Scheltjens (Rijksmuseum Amsterdam): Open Rijksmuseum Data: challenges and opportunities
16:30-17:15 Nanne van Noord (Universiteit Tilburg): Learning visual representations of style
17:15-18:30 Reception

Understanding the perils of Linked Data through the history of data modeling

Seth van Hooland

Based on concrete examples from the cultural heritage world, this talk illustrates how different data models were developed, each one offering its own possibilities and limits. Putting RDF in this larger context helps to develop a more critical understanding of Linked Data.

Seth van Hooland studies document and records management strategies and enjoys cleaning up dirty metadata. He is the deputy head of the Information and Communication Science department and responsible for the Master in Information Science at ULB.


The Replica Project : Navigating Iconographic Collections at Scale

Benoit Seguin

In recent years, museums and institutions have pushed fBenoit Seguin
or a global digitization effort of their collections. Millions of artefacts (paintings, engravings, sketches, old photographs, …) are now in digital photographic format. Furthermore, through the IIIF standards, a significant portion of these images are now available online in an easily accessible manner. On the other side of the spectrum, the recent advances in Machine Learning are allowing computers to tackle more and more complex visual tasks.

The combination of more data and better technology opens new opportunities for Art Historians to navigate these collections. While most of the time search queries are solely based on textual information (metadata, tags, …), we focus on visual similarities i.e working directly with the images.

I will quickly report on our digitization effort with the processing of the fototeca of the Foundazione Giorgio Cini, where more than 300’000 images are already digitized. Then I will explain how we worked with Art Historians to learn a specific image similarity function by leveraging the concept of Visual Links between images. Finally, I will present our work on the user interfaces leveraging such a metric, allowing new ways to explore large collections of images, and how can users further refine the tool.

Benoit Seguin is a PhD Candidate at the Digital Humanities Laboratory of EPFL. Prior to starting his PhD, Benoit got his Diplôme d’Ingénieur at Ecole Polytechnique ParisTech and a MSc in Computer Science from EPFL. Before applying Image Processing and Computer Vision to Digital Humanities, he was using similar techniques in Circuit Manufacturing, Robotics, and Medical Imaging. The main topic of his PhD is to use Machine Learning to explore the next generation of tools for Digital Art History.


Open Rijksmuseum Data : challenges and opportunities

Saskia Scheltjens [slides]

The Rijksmuseum, the National Museum of Art and History of the Netherlands, has been on an open data journey for quite some time now. From opening up a limited collection of images and collection data with a CCBY license in 2011 to full blown sharing of all of its high resolution collection images and metadata with a CC0 license in 2013, the Rijksmuseum is hailed as one of the champions of the open content movement in the international museum world. What was the reasoning behind this bold move? And does that reasoning still holds up today? Precisely what kind of content do we speak of, and what is the impact of the size of the collection and the datasets that are involved. This presentation will highlight some of the decisions that were made in the past, and try to sketch some of the current challenges and new developments the museum is working on.

Saskia Scheltjens (1970) studied Dutch and English Literature and Linguistics, and Information and Library Science at the University of Antwerp (Belgium). She has worked as Head of the Library at the Museum of Contemporary Art in Ostend, and was responsible for a very large reorganisation and the formation of a new Faculty Library of Arts & Philosophy at Ghent University. In 2016 she was asked to set up a new department called Research Services at the Rijksmuseum in Amsterdam. Together with her team, she is responsible for the collection information and data, be it analog or digital, be it the metadata of object collections, documentation, the museum library, object archives and research data. Saskia is fascinated by the interdisciplinary possibilities of digital humanities research and a strong advocate for open data within the digital heritage world.


Learning visual representations of style

Nanne van Noord [slides part 1slides part 2]

An artist’s style is reflected in their artworks, independent from what is depicted. Two artworks created by the same artist that depict two vastly different scenes (e.g., a beach scene and a forest scene) both reflect their style. Stylistic characteristics of an artwork can be used by experts, and sometimes even laymen, to identify the artist that created the artwork. The ability to recognize styles and relate these to artists is associated with connoisseurship. Connoisseurship is essential in the tasks of authentication and restoration of artworks, because both tasks require detailed knowledge of stylistic characteristics of the artist. In this talk I will cover a number of studies we performed aimed at realising connoisseurship in a computer, for both recognition and production tasks.

Nanne van Noord is a Postdoc working on the SEMIA (The Sensory Moving Image Archive) project at the Informatics Institute, University of Amsterdam. Previously he was a PhD candidate working on the REVIGO (REassessing VIncent van GOgh) project at Tilburg University. His research interests include deep learning, image processing, and digital cultural heritage.


International Image Interoperability Framework (IIIF). Sharing high-resolution images across institutional boundaries

Roxanne Wyns [slides]

IIIF or the International Image Interoperability Framework is a community-developed framework for sharing high-resolution images in an efficient and standardized way across institutional boundaries. Using an IIIF manifest URL, a researcher can simply pull the images and related contextual information such as the structure of a complex object or document, metadata and rights information into any IIIF compliant viewer such as the Mirador viewer. Simply put, a researcher can access a digital resource from the British Library and from the KU Leuven Libraries in a single viewer for research, while allowing the institutions to exert control over the quality and context of the resources offered. KU Leuven implemented IIIF in 2015 in the framework of the idemdatabase.org project and has since been using it in a number of Digital Humanities projects with a focus on high-resolution image databases for research. By now the IIIF community has grown considerably with institutions such as the ‘The J. Paul Getty Trust’ and the ’Bibliotheca Vaticana’ implementing it as a standard and providing swift and standardized access to thousands of resources. Its potential has however not reached its limits with ongoing work on aspects such IIIF resource discovery and harvesting of manifest URLs, annotation functions and an extension to include audio-visual material. The presentation will introduce IIIF and its concepts, highlight projects and viewers, and give an in-depth view of its current and future application options.

Roxanne Wyns studied History of Art and Archaeology at the Free University of Brussels (VUB). Since 2009 she worked on several European projects, specialising in standards, multilingual thesaurus management, and data interoperability and aggregation processes. At LIBIS – KU Leuven she supports KU Leuven and its partners in realizing their digital strategy. She is involved in several research infrastructure projects and supports the ‘Services for researchers’ in the framework of Research Data Management (RDM). Roxanne is actively involved in the IIIF community and is a member of the Dariah-EU Scientific Advisory Board.