🔥Released OpenThaiGPT 7b <1.0.0-beta> (16/08/23)

🇹🇭 OpenThaiGPT 1.0.0-beta (16 August 2023)

🇹🇭 OpenThaiGPT Version 1.0.0-beta is a Thai language 7B-parameter LLaMA v2 Chat model finetuned to follow Thai translated instructions and extend more than 24,554 most popular Thai words vocabularies into LLM's dictionary for turbo speed.

Web Demo:

Colab Demo:

Change Logs

🇹🇭 Version 1.0.0-beta (Llama v2 + 24,554 Thai word extension)

Release date: 16 August 2023

🇹🇭 OpenThaiGPT Version 1.0.0-beta is a Thai language 7B-parameter LLaMA v2 Chat model finetuned to follow Thai translated instructions and extend 24,554 Thai words vocabularies for turbo speed.

License

Source Code: License Apache Software License 2.0. Weight: Research and Commercial uses.

Code and Weight

Finetune Code: https://github.com/OpenThaiGPT/openthaigpt-finetune-010beta Inference Code: https://github.com/OpenThaiGPT/openthaigpt Weight: https://huggingface.co/openthaigpt/openthaigpt-1.0.0-beta-7b-chat

Authors

  • Kobkrit Viriyayudhakorn (kobkrit@aieat.or.th)

  • Sumeth Yuenyong (sumeth.yue@mahidol.edu)

  • Prachya Boonkwan (prachya.boonkwan@nectec.or.th, kaamanita@gmail.com)

  • Thaweewat Rugsujarit (thaweewr@scg.com)

  • Jillaphat Jaroenkantasima (autsadang41@gmail.com)

  • Norapat Buppodom (new@norapat.com)

  • Koravich Sangkaew (kwankoravich@gmail.com)

  • Peerawat Rojratchadakorn (peerawat.roj@gmail.com)

  • Surapon Nonesung (nonesungsurapon@gmail.com)

  • Chanon Utupon (chanon.utupon@gmail.com)

  • Sadhis Wongprayoon (sadhis.tae@gmail.com)

  • Nucharee Thongthungwong (nuchhub@hotmail.com)

  • Chawakorn Phiantham (mondcha1507@gmail.com)

  • Patteera Triamamornwooth (patt.patteera@gmail.com)

  • Nattarika Juntarapaoraya (natt.juntara@gmail.com)

  • Kriangkrai Saetan (kraitan.ss21@gmail.com)

  • Pitikorn Khlaisamniang (pitikorn32@gmail.com)

  • Teerapol Saengsukhiran (winroom@gmail.com)

  • Phasin Aumwong (phasin03895@gmail.com)

---

🇹🇭 Version 1.0.0-alpha (Facebook LLama V2 Model)

Release date: 3 August 2023

🇹🇭 OpenThaiGPT Version 1.0.0-alpha is the first Thai implementation of a 7B-parameter LLaMA v2 Chat model finetuned to follow Thai translated instructions and makes use of the Huggingface LLaMA implementation.

Changes

(1) Using Facebook LLama v2 model 7b chat as a base model which is pretrained on over 2 trillion token. (2) Context Length is upgrade from 2048 token to 4096 token (3) Allow research and commerical use.

License

Source Code: License Apache Software License 2.0. Weight: Research and commercial uses.

Code and Weight

Colab Demo: https://colab.research.google.com/drive/1kDQidCtY9lDpk49i7P3JjLAcJM04lawu?usp=sharing Finetune Code: https://github.com/OpenThaiGPT/openthaigpt-finetune-010beta (Same code as OpenThaiGPT 0.1.0-beta) Inference Library: https://github.com/OpenThaiGPT/openthaigpt Weight (Lora Adapter): https://huggingface.co/openthaigpt/openthaigpt-1.0.0-alpha-7b-chat Weight (Huggingface Checkpoint): https://huggingface.co/openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ckpt-hf

Authors

  • Kobkrit Viriyayudhakorn (kobkrit@aieat.or.th)

  • Sumeth Yuenyong (sumeth.yue@mahidol.edu)

  • Thaweewat Rugsujarit (thaweewr@scg.com)

  • Jillaphat Jaroenkantasima (autsadang41@gmail.com)

  • Norapat Buppodom (new@norapat.com)

  • Koravich Sangkaew (kwankoravich@gmail.com)

  • Peerawat Rojratchadakorn (peerawat.roj@gmail.com)

  • Surapon Nonesung (nonesungsurapon@gmail.com)

  • Chanon Utupon (chanon.utupon@gmail.com)

  • Sadhis Wongprayoon (sadhis.tae@gmail.com)

  • Nucharee Thongthungwong (nuchhub@hotmail.com)

  • Chawakorn Phiantham (mondcha1507@gmail.com)

  • Patteera Triamamornwooth (patt.patteera@gmail.com)

  • Nattarika Juntarapaoraya (natt.juntara@gmail.com)

  • Kriangkrai Saetan (kraitan.ss21@gmail.com)

  • Pitikorn Khlaisamniang (pitikorn32@gmail.com)

  • Teerapol Saengsukhiran (winroom@gmail.com)

  • Phasin Aumwong (phasin03895@gmail.com)

---

Version 0.1.0-beta (Facebook LLama Model)

Release date: 16 May 2023

OpenThaiGPT Version 0.1.0-beta is a 7B-parameter LLaMA model finetuned to follow Thai translated instructions below and makes use of the Huggingface LLaMA implementation.

Statistics

Number of parameters: 7B Dimension: 4096 Context Length: 2048 n heads: 32 n layers: 32 n tokens: 1T

License

Source Code: License Apache Software License 2.0. Weight: For research use only (due to the Facebook LLama's Weight LICENSE). Note that: A commercial use license for OpenThaiGPT 0.1.0 weight will be released later soon!

Code and Weight

Finetune Code: https://github.com/OpenThaiGPT/openthaigpt-finetune-010beta Inference Library: https://github.com/OpenThaiGPT/openthaigpt Weight (Lora Adapter): https://huggingface.co/kobkrit/openthaigpt-0.1.0-beta

Authors

Kobkrit Viriyayudhakorn (kobkrit@aieat.or.th), Sumeth Yuenyong (sumeth.yue@mahidol.edu) and Thaweewat Rugsujarit (thaweewr@scg.com).

Trained Datasets

---

Version 0.1.0-alpha (ByT5-XL Model)

Release date: 24 April 2023 PoC Testing Website: https://colab.research.google.com/drive/1Uds0ioOZSZrJ9m2FgW3DHlqVRFNHVRtu#scrollTo=qPJIpwuz4ltF Model and Weight: https://huggingface.co/kobkrit/openthaigpt-0.1.0-alpha PIP Installation Page: https://pypi.org/project/openthaigpt/ Code Example: https://colab.research.google.com/drive/1Uds0ioOZSZrJ9m2FgW3DHlqVRFNHVRtu#scrollTo=qPJIpwuz4ltF ----

OpenThaiGPT version 0.1.0-alpha

Thai First 3 billion params models

  • First Thai Byte-Level Text-to-Text Transfer Transformer

  • Support Instruction following

    • Translation to Thai

    • Explanation

    • Paraphase

  • Zero-shot and Few-shot Learning

  • Pretraining Model: ByT5-XL (3.74 billion params)

  • InstructDataset: 50,000 Thai SelfInstruct

  • RLHF: None

  • Developer: Sumeth Yuenyong, Kobkrit Viriyayudhakorn (kobkrit@iapp.co.th)

PoC Version 0.0.4 (The Fourth PoC Version)

Release date: 12 March 2023 PoC Testing Website: https://colab.research.google.com/drive/13yLIifBRDQp82QO4ICs_aEvz0N8tqVPm?usp=sharin Model and Weight: https://huggingface.co/kobkrit/openthaigpt-gpt2-instructgpt-poc-0.0.4 PIP Installation Page: https://pypi.org/project/openthaigpt/ Code Example: https://github.com/OpenThaiGPT/openthaigpt-example ----

OpenThaiGPT version 0.0.4

The Fourth PoC Model

  • ตอบคำถามได้ลงรายละเอียดมากขึ้น และตอบคำถามได้ดีขึ้นกว่า 0.0.3 เป็นส่วนมาก

  • Pretraining Model: GPT-2 Thai-base

  • InstructDataset: 300,000 Pantip + 5,000 Wiki QA => 12,920 Thai InstructGPT

  • RLHF: None

  • Developer: Kobkrit Viriyayudhakorn (kobkrit@iapp.co.th)

PoC Version 0.0.3 (The Third PoC Version)

Release date: 28 February 2023 Model and Weight: https://huggingface.co/kobkrit/openthaigpt-gpt2-instructgpt-poc-0.0.3 PIP Installation Page: https://pypi.org/project/openthaigpt/ Code Example: https://github.com/OpenThaiGPT/openthaigpt-example ----

OpenThaiGPT version 0.0.3

The Third PoC Model

  • Pretraining Model: GPT-2 Thai-base

  • InstructDataset: 300,000 Pantip + 5,000 Wiki QA => 7,000 Thai InstructGPT

  • RLHF: None

  • Developer: Kobkrit Viriyayudhakorn (kobkrit@iapp.co.th)

PoC Version 0.0.2 (The Second PoC Version)

Release date: 27 February 2023 Model and Weight: https://huggingface.co/kobkrit/openthaigpt-gpt2-instructgpt-poc-0.0.2 PIP Installation Page: {Coming Soon} Colab Example: {Coming Soon} ----

OpenThaiGPT version 0.0.2

The Second PoC Model

  • Pretraining Model: GPT-2 Thai-base

  • InstructDataset: 7,000 Thai InstructGPT

  • RLHF: None

Developer: Kobkrit Viriyayudhakorn (kobkrit@iapp.co.th)

PoC Version 0.0.1 (Very First PoC Version)

Release date: 20 February 2023 Model and Weight: openthaigpt-gpt2-pantipwiki-poc PIP Installation Page: {Coming Soon} Colab Example: {Coming Soon} ----

The Very First PoC Model

  • Pretraining Model: GPT-2 Thai-base

  • InstructDataset: 298,678 QA Pairs getting from 70,000 Pantip katoos + Wikipedia QA by iApp

  • RLHF: None

  • Developer: Kobkrit Viriyayudhakorn (kobkrit@iapp.co.th)

Last updated