Model Card

๋ชจ๋ธ ๊ฐœ์š”

์ด ๋ชจ๋ธ์€ **bert-base-uncased**๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํŒŒ์ธํŠœ๋‹ํ•œ ํ…์ŠคํŠธ ๋ถ„๋ฅ˜ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ๋Š” AG News๋ฅผ ์‚ฌ์šฉํ•˜์˜€์œผ๋ฉฐ, ์ด 4๊ฐœ์˜ ์นดํ…Œ๊ณ ๋ฆฌ(์„ธ๊ณ„๋‰ด์Šค, ์Šคํฌ์ธ , ๋น„์ฆˆ๋‹ˆ์Šค, ๊ณผํ•™/๊ธฐ์ˆ )๋กœ ๋ถ„๋ฅ˜ํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

  • ๋ฒ ์ด์Šค ๋ชจ๋ธ: bert-base-uncased

  • ํŒŒ์ธํŠœ๋‹ ๋ฐ์ดํ„ฐ: AG News

  • ํƒœ์Šคํฌ(Task): ํ…์ŠคํŠธ ๋ถ„๋ฅ˜ (4-class classification)

  • ๋ผ๋ฒจ(Label) ์ •์˜:

    • 0: ์„ธ๊ณ„๋‰ด์Šค (World News)
    • 1: ์Šคํฌ์ธ  (Sports)
    • 2: ๋น„์ฆˆ๋‹ˆ์Šค (Business)
    • 3: ๊ณผํ•™/๊ธฐ์ˆ  (Sci/Tech)

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch

# ๋ชจ๋ธ ๋กœ๋“œ
model_name = "blockenters/bert-based-uncased-agnews4-v01"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)

# ์ž…๋ ฅ ์˜ˆ์‹œ
text = "NASA launches a new satellite into space"
inputs = tokenizer(text, return_tensors="pt")

# ์ถ”๋ก 
with torch.no_grad():
    outputs = model(**inputs)
    predictions = torch.argmax(outputs.logits, dim=-1)

print("Predicted label:", predictions.item())

์„ฑ๋Šฅ

  • ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹: AG News test set

  • ๋ฉ”ํŠธ๋ฆญ: Accuracy

  • ๊ฒฐ๊ณผ (์˜ˆ์‹œ ๊ฐ’, ์ง์ ‘ ์ธก์ •ํ•ด์„œ ์ฑ„์›Œ์ฃผ์„ธ์š”):

    • Accuracy: 0.86
    • F1-score (weighted): 0.86

ํ™œ์šฉ ์‚ฌ๋ก€

  • ๋‰ด์Šค ๊ธฐ์‚ฌ ์ž๋™ ๋ถ„๋ฅ˜
  • ๋‰ด์Šค ์ถ”์ฒœ ์‹œ์Šคํ…œ์˜ ์ „์ฒ˜๋ฆฌ ๋‹จ๊ณ„
  • ํ† ํ”ฝ ๊ธฐ๋ฐ˜ ํ…์ŠคํŠธ ๋ถ„์„

์ œํ•œ ์‚ฌํ•ญ

  • ์˜์–ด ๋ฐ์ดํ„ฐ์…‹(AG News)๋กœ ํ•™์Šตํ–ˆ๊ธฐ ๋•Œ๋ฌธ์— ์˜์–ด ํ…์ŠคํŠธ ๋ถ„๋ฅ˜์— ์ตœ์ ํ™”๋จ
  • ๋‹ค๋ฅธ ์–ธ์–ด ํ…์ŠคํŠธ์—๋Š” ์„ฑ๋Šฅ์ด ๋–จ์–ด์งˆ ์ˆ˜ ์žˆ์Œ
  • ์ตœ์‹  ๋‰ด์Šค ์ด๋ฒคํŠธ์— ๋Œ€ํ•œ ํ•™์Šต์€ ํฌํ•จ๋˜์–ด ์žˆ์ง€ ์•Š์Œ

๋ผ์ด์„ ์Šค

  • ๋ฒ ์ด์Šค ๋ชจ๋ธ(bert-base-uncased)์˜ ๋ผ์ด์„ ์Šค๋Š” Apache 2.0์ž…๋‹ˆ๋‹ค.
  • AG News ๋ฐ์ดํ„ฐ์…‹์€ open dataset์œผ๋กœ ๊ณต๊ฐœ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค.
Downloads last month
6
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support