jayralencar / pictogram-to-sense

Mapping pictograph symbols to WordNet word-senses

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Pictogram to Sense: mapping pictographic symbols to WordNet synsets

Pictogram is the term used by the Augmentative and Alternative Communication (AAC) community for an image with a label that represents a place, person, action, object and animal. AAC systems like the shown below allow message construction and communication by arranging pictograms in sequence.

image

Every day the need to use artificial intelligence in tools of this type grows. An example is PictoBERT, an adaptation of BERT language model to perform pictogram prediction. PictoBERT use word-sense language modeling to perform pictogram prediction. This way, each pictogram must be associated with a word-sense. Besides, assosiating pictograms to word-senses allow they to be used in other natural language processing applications related to AAC (e.g., text expansion).

This project aims to provide AAC developers with a tool for facilitating the mapping between pictograms and WordNet word-senses. For this, we use CLIP (Constrative Language-Image Pre-Training), which learns visual concepts from natural language supervision, using image+text pairs for training. The procedure consists on searching for the synsets of the pictogram label in WordNet. CLIP encodes the synsets definitions and the pictogram image and calculates a cosine similarity between each encoded definition and the image. The similarity scores are then scaled by a temperature parameter τ, and normalized into a probability distribution via a softmax. The definition with the highest value after softmax is chosen for the pictogram, as well as its synset. To improve the quality of predictions made by CLIP for this task, we trained the model for a few more epochs using ARASAAC pictograms and their definitions as a dataset.

image

image

For evaluating our proposal, we attributed synsets to 900 of the pictograms from Mulberry Symbols, and calculated how accurate is CLIP on selecting the correct pictogram. We evaluated both versions of CLIP, the original and the fine-tuned. The results demonstrate that the fine-tuned models have a better performance than the original CLIP. Besides, the accuracy score acquired (70%) demonstrates the proposal's potential.

image

Usage

Execute the notebooks in sequence:

  1. Dataset Creation Notebook
  2. Fine-tuning Notebook
  3. Evaluating Notebook

About

Mapping pictograph symbols to WordNet word-senses

License:MIT License


Languages

Language:Jupyter Notebook 100.0%