
Overview
VARCO-LLM is NCSOFT’s large language model, which can be applied to develop various NLP-based AI services such as Q&A, chatbot, summarization, information extraction etc. VARCO-LLM, trained with public pre-training data and internally constructed high-quality Korean data, boasts the highest performance among the Korean LLMs of similar sizes that have been released to date (see https://ncsoft.github.io/ncresearch/ for evaluation results). Our models will continue to be updated and we will also release LLMs that support multiple languages or are fined-tuned to specific tasks. As VARCO-LLM is currently in beta service (29 Aug to 10 Sep 2023), usage fees will not be charged temporally for this period. For inquiries regarding further performance improvement or collaboration for service applications, please contact us via email (varco_llm@ncsoft.com ).
Highlights
- The VARCO LLM KO/EN-13B-IST is a bilingual instruction-tuned model that is trained with the Korean/English pre-training data and the instruction dataset, both constructed by NCSOFT.
- This model surpasses the VARCO LLM KO-13B-IST model in Korean/English text generation tasks, enabling it to be utilized globally in Korean/English speaking countries with one model.
- It is trained with a 10x larger instruction dataset than the VARCO LLM KO-13B-IST model, thus it can effectively respond to many more diverse types of prompts.
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Pricing
Vendor refund policy
This product is offered for free. If there are any questions, please contact us for further clarifications.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker model
An Amazon SageMaker model package is a pre-trained machine learning model ready to use without additional training. Use the model package to create a model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
Safety enhanced model update
Additional details
Inputs
- Summary
Model accepts JSON requests. You can check examples and fields descriptions.
- Input MIME type
- application/json
Input data descriptions
The following table describes supported input data fields for real-time inference and batch transform.
Field name | Description | Constraints | Required |
|---|---|---|---|
text | The sequence used as a prompt for the generation or as model inputs to the encoder. | Type: FreeText | Yes |
request_output_len | The maximum numbers of tokens to generate, ignoring the number of tokens in the prompt. | Default value: 1024
Type: Integer
Maximum: 2048 | No |
repetition_penalty | The parameter for repetition penalty. 1.0 means no penalty. | Default value: 1.05
Type: Continuous | No |
temperature | The value used to modulate the next token probabilities. | Default value: 0.1
Type: Continuous | No |
top_k | The number of highest probability vocabulary tokens to keep for top-k-filtering. | Default value: 50
Type: Integer | No |
top_p | If set to float < 1, only the smallest set of most probable tokens with probabilities that add up to top_p or higher are kept for generation.
| Default value: 1.0
Type: Continuous | No |
Resources
Vendor resources
Support
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products



