PascalSentenceDataset
This program is utility to download pascal sentence dataset using Python3 with a virtual environment set up by Poetry. All credits go to: https://github.com/rupy
Installation
You can install by "git clone" command.
https://github.com/gorjanradevski/PascalSentenceDataset.git
Dependency
You must install some python libraries. Use poetry install
once Poetry has been set up.
Usage
To download dataset, just run program as follow:
poetry run python pascal_sentence_dataset.py
You can also write code like this:
# import
from pascal_sentence_dataset import PascalSentenceDataSet
# create instance
dataset = PascalSentenceDataSet()
# download images
dataset.download_images()
# download sentences
dataset.download_sentences()
# create correspondence data by dataset
dataset.create_correspondence_data()
That's it!
Correspondence data is the csv data to correspond data id to image data.
Additional information for Nakayama lab members
Our lab created Japanese translation of Pascal Sentence Dataset. Translation class is the utility to use parallel translation data, "pascal_sentence_numbers.csv". You can get text files of two languages by the class. To use the class, you have to install depencent libraries as follow:
mecab-python 0.996
To use mecab-python, you have to install MeCab in addition.
Usage
To create Japanese & English parallel translation data, just run program as follow:
python pascal_sentence_dataset.py
You can also write code like this:
# import
from translation import Translation
# put parallel translation data somewhere in advance
csv_file = 'translations/pascal_sentence_numbers.csv'
# initialize instance
ps = Translation(csv_file)
# create text data from csv file
ps.read_csv_and_save_as_txt()
# create wakati-gaki text data (Japanese text data separated by space between each word)
ps.wakati()