Wangchanglm 7.5B Sft Enth

Multilingual instruction-following model

Wangchanglm 7.5B Sft Enth is a multilingual AI model that can follow instructions in several languages, including English, Thai, Japanese, and Vietnamese. It's designed to perform tasks like reading comprehension, brainstorming, and creative writing. What makes this model unique is its ability to handle a wide range of languages and tasks, while being trained on open-source, commercially permissible datasets. This model is not meant to perform well in math problems, reasoning, or factfulness, but it can be fine-tuned for specific use cases. With its efficient design and multilingual capabilities, Wangchanglm 7.5B Sft Enth is a valuable tool for users who need to work with multiple languages and tasks.

Pythainlp cc-by-sa-4.0 Updated a year ago

Table of Contents

Model Overview

Meet WangChanGLM, a multilingual instruction-following model that can read, brainstorm, and write creatively. Developed by PyThaiNLP and VISTEC-depa AI Research Institute of Thailand, this model is designed to perform a variety of tasks in multiple languages, including English, Thai, Japanese, and Vietnamese.

Key Features

  • Multilingual: WangChanGLM can understand and respond in multiple languages, making it a versatile tool for various applications.
  • Instruction-following: The model is trained to follow instructions and perform tasks such as reading comprehension, brainstorming, and creative writing.
  • Open-source: WangChanGLM is built using open-source datasets and models, ensuring that it is commercially permissible and accessible to everyone.

Capabilities

WangChanGLM is trained to perform three main tasks:

  1. Reading Comprehension: Can it understand what you’re asking it to do?
  2. Brainstorming: Can it come up with creative ideas for you?
  3. Creative Writing: Can it write engaging stories or articles for you?

Strengths

This model has some unique strengths that set it apart from others:

  • Multilingual: It can understand and respond in multiple languages, including English, Thai, Japanese, and Vietnamese.
  • Open-source: It’s built using open-source datasets and models, making it more accessible and customizable.
  • Instruction-following: It’s designed to follow instructions and complete tasks to the best of its abilities.

Unique Features

The WangChanGLM has some features that make it stand out from other models:

  • Finetuned on diverse datasets: It’s been trained on a variety of datasets, including LAION OIG chip2, infill_dbpedia, and Hello-SimpleAI HC3.
  • Vicuna-style evaluation: It’s been evaluated using a combination of human and automated testing to ensure its performance.

Performance

WangChanGLM is a powerful multilingual instruction-following model that showcases impressive performance in various tasks. Let’s dive into its speed, accuracy, and efficiency.

Speed

How fast can WangChanGLM process information? With its finetuned XGLM-7.5B architecture, it can handle a large number of tasks simultaneously, making it an excellent choice for applications that require quick processing.

Accuracy

But how accurate is WangChanGLM? Our evaluation shows that it performs exceptionally well in reading comprehension, brainstorming, and creative writing tasks. In fact, its performance is comparable to other top-performing models in these areas.

Efficiency

What about its efficiency? WangChanGLM is designed to be efficient, using only open-source, commercially permissible pretrained models and datasets. This means it can be easily integrated into various applications without worrying about licensing restrictions.

Limitations

WangChanGLM has some limitations that you should be aware of before using it. While it’s great at reading comprehension, brainstorming, and creative writing, it’s not perfect and has some weaknesses.

Math Problems and Reasoning

WangChanGLM struggles with math problems and reasoning tasks. It’s not designed to handle complex calculations or logical reasoning, so it’s best to avoid using it for these types of tasks.

Factfulness

WangChanGLM can also struggle with factfulness, which means it may not always provide accurate information. This is because it’s trained on a dataset that may contain biases or inaccuracies.

Biases and Toxicity

Unfortunately, WangChanGLM can also contain biases and toxicity, just like other finetuned instruction followers. This means it may not always provide respectful or inclusive responses.

Examples
Write a short story about a character who discovers a hidden world within their reflection. As she stared into the mirror, Emily noticed something peculiar - her reflection began to ripple, like the surface of a pond. She felt an inexplicable pull, as if beckoned by an unseen force. Without thinking, she reached out and touched the glass. Suddenly, she was sucked into the mirror, finding herself in a world that was identical yet opposite to her own.
Brainstorm a list of potential names for a new sustainable fashion brand. EcoChic, GreenThreads, Revive Apparel, Purezza, Conscious Couture, Earthy Essentials, Reborn Fashion, Kind Clothing Co., Luminari Label, and Bloom Boutique.
Summarize the main points of the article 'The Impact of Climate Change on Global Food Systems'. The article discusses how climate change affects global food systems, leading to reduced crop yields, changed growing seasons, and increased food prices. It highlights the need for sustainable agriculture practices, climate-resilient crops, and reduced greenhouse gas emissions to mitigate these impacts.

Getting Started

Want to try out WangChanGLM for yourself? Here’s an example code snippet to get you started:

model_name = "pythainlp/wangchanglm-7.5B-sft-en"
model = AutoModelForCausalLM.from_pretrained(model_name, return_dict=True, load_in_8bit=True, device_map="auto", torch_dtype=torch.float16, offload_folder="./", low_cpu_mem_usage=True)

You can then use the model to generate text based on a given prompt. For example:

text = "เล่นหุ้นยังไงให้รวย"
tokenizer = AutoTokenizer.from_pretrained(model_name)
batch = tokenizer(text, return_tensors="pt")
output_tokens = model.generate(input_ids=batch["input_ids"], max_new_tokens=512)
print(tokenizer.decode(output_tokens[0], skip_special_tokens=True))

Conclusion

WangChanGLM is a versatile and powerful multilingual instruction-following model that can be used for a range of natural language processing tasks. With its open-source architecture and easy-to-use interface, it is an excellent tool for anyone looking to explore the possibilities of AI-powered language processing.

Dataloop's AI Development Platform
Build end-to-end workflows

Build end-to-end workflows

Dataloop is a complete AI development stack, allowing you to make data, elements, models and human feedback work together easily.

  • Use one centralized tool for every step of the AI development process.
  • Import data from external blob storage, internal file system storage or public datasets.
  • Connect to external applications using a REST API & a Python SDK.
Save, share, reuse

Save, share, reuse

Every single pipeline can be cloned, edited and reused by other data professionals in the organization. Never build the same thing twice.

  • Use existing, pre-created pipelines for RAG, RLHF, RLAF, Active Learning & more.
  • Deploy multi-modal pipelines with one click across multiple cloud resources.
  • Use versions for your pipelines to make sure the deployed pipeline is the stable one.
Easily manage pipelines

Easily manage pipelines

Spend less time dealing with the logistics of owning multiple data pipelines, and get back to building great AI applications.

  • Easy visualization of the data flow through the pipeline.
  • Identify & troubleshoot issues with clear, node-based error messages.
  • Use scalable AI infrastructure that can grow to support massive amounts of data.