File size: 3,094 Bytes
a98b3ce 534880f 3c4bf60 534880f 3c4bf60 534880f 3c4bf60 534880f 2b8bf91 534880f 56e9fdd 534880f 3c4bf60 534880f 3c4bf60 534880f 56e9fdd 534880f 56e9fdd 534880f 3c4bf60 534880f 56e9fdd 534880f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 |
---
license: apache-2.0
language:
- ko
- en
pipeline_tag: visual-question-answering
tags:
- text2text-generation
base_model: google/deplot
---
# **ko-deplot**
ko-deplot is a korean Visual-QA model based on the Google's Pix2Struct architecture. It was fine-tuned from [Deplot](https://huggingface.co./google/deplot), using korean chart image-text pairs.
ko-deplot์ Google์ Pix2Struct ๊ตฌ์กฐ๋ฅผ ๊ธฐ๋ฐ์ผ๋ก ํ ํ๊ตญ์ด Visual-QA ๋ชจ๋ธ์
๋๋ค. [Deplot](https://huggingface.co./google/deplot) ๋ชจ๋ธ์ ํ๊ตญ์ด ์ฐจํธ ์ด๋ฏธ์ง-ํ
์คํธ ์ ๋ฐ์ดํฐ์
์ ์ด์ฉํ์ฌ ํ์ธํ๋ํ์์ต๋๋ค.
- **Developed by:** [NUUA](https://www.nuua.ai/en/)
- **Model type:** Visual Question Answering
- **License:** apache-2.0
- **Finetuned from model:** [google/deplot](https://huggingface.co./google/deplot)
# **Model Usage**
You can run a prediction by querying an input image together with a question as follows:
์๋์ ์ฝ๋๋ฅผ ์ด์ฉํ์ฌ ๋ชจ๋ธ ์ถ๋ก ์ ํ ์ ์์ต๋๋ค:
```python
from transformers import Pix2StructProcessor, Pix2StructForConditionalGeneration
from PIL import Image
processor = Pix2StructProcessor.from_pretrained('nuua/ko-deplot')
model = Pix2StructForConditionalGeneration.from_pretrained('nuua/ko-deplot')
IMAGE_PATH = "LOCAL_PATH_TO_IMAGE"
image = Image.open(IMAGE_PATH)
inputs = processor(images=image, text="Generate underlying data table of the figure below:", return_tensors="pt")
predictions = model.generate(**inputs, max_new_tokens=512)
print(processor.decode(predictions[0], skip_special_tokens=True))
```
# **Tokenizer Details**
The model's tokenizer vocab was extended from 50,344 to 65,536 tokens using the following:
- Complete Korean Jamo
- [Additional Korean Jamo](http://koreantypography.org/wp-content/uploads/2016/02/kst_12_7_2_06.pdf)
- Ko-Electra tokens
๋ชจ๋ธ์ tokenizer vocab์ 50344๊ฐ์์ 65536๊ฐ๋ก ์๋๋ฅผ ์ด์ฉํ์ฌ ํ์ฅ์ํจ ํ ํ์ต์ ์งํํ์์ต๋๋ค:
- ์์ฑํ ํ๊ธ ์๋ชจ
- [์ถ๊ฐ ์์ฑํ ํ๊ธ ์๋ชจ](http://koreantypography.org/wp-content/uploads/2016/02/kst_12_7_2_06.pdf)
- Ko-Electra ํ๊ธ ํ ํฐ
# **Training Details**
## Training Data
Synthetic chart data from three libraries were used:
์ธ ๊ฐ์ ๋ผ์ด๋ธ๋ฌ๋ฆฌ์์ ํฉ์ฑ ์ฐจํธ ๋ฐ์ดํฐ๋ฅผ ์์ฑํ์ฌ ์ฌ์ฉํ์์ต๋๋ค:
- [GenPlot](https://github.com/brendanartley/genplot)
- [Chart.js](https://github.com/chartjs/Chart.js)
- [Plotly](https://github.com/plotly/plotly.py)
## Training Procedure
The model was first exposed to a short warmup stage, following its [original paper](https://arxiv.org/pdf/2210.03347.pdf). It was then trained using the chart data for 50,000 steps.
ํ์ต์ ์ํด ์ฒ์ ์งง์ "warmup" ๋จ๊ณ๋ฅผ ๊ฑฐ์ณ ํ๊ธ์ ํ์ต์ํจ ํ 50,000 ์คํ
๋์ ์ฐจํธ ๋ฐ์ดํฐ๋ฅผ ํ์ต์์ผฐ์ต๋๋ค.
# **Technical Specifications**
## Hardware
ko-deplot was trained by using A100 80G.
A100 80G GPU๋ฅผ ์ด์ฉํ์ฌ ํ์ตํ์์ต๋๋ค.
# **Contact**
Any questions and suggestions, please use the discussion tab. If you want to contact us directly, email [email protected]. |