kor-smishing-xlmroberta

πŸ“Œ Model Overview

kor-smishing-xlmrobertaλŠ”
XLM-RoBERTa 기반의 ν•œκ΅­μ–΄ μŠ€λ―Έμ‹±(Smishing) 탐지 λͺ¨λΈμž…λ‹ˆλ‹€.

ν•œκ΅­μ–΄ SMS 및 λ©”μ‹ μ € ν…μŠ€νŠΈλ₯Ό μž…λ ₯으둜 λ°›μ•„
ν•΄λ‹Ή λ©”μ‹œμ§€κ°€ μŠ€λ―Έμ‹±(사기)인지 μ—¬λΆ€λ₯Ό 이진 λΆ„λ₯˜ν•˜λ„둝 νŒŒμΈνŠœλ‹λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

λ³Έ λͺ¨λΈμ€ κ³Όλ„ν•œ κ·œμΉ™(rule) 기반 νœ΄λ¦¬μŠ€ν‹±μ— μ˜μ‘΄ν•˜μ§€ μ•Šκ³ ,
λ”₯λŸ¬λ‹ λͺ¨λΈ λ‹¨λ…μœΌλ‘œλ„ **높은 정밀도(Precision)**λ₯Ό λ‹¬μ„±ν•˜λŠ” 것을 λͺ©ν‘œλ‘œ μ„€κ³„λ˜μ—ˆμŠ΅λ‹ˆλ‹€.


🧠 Model Architecture

  • Base Model: xlm-roberta-base
  • Model Type: XLMRobertaForSequenceClassification
  • Task: Binary Text Classification
  • Output Labels
    • LABEL_0: 정상 (HAM)
    • LABEL_1: μŠ€λ―Έμ‹± (PHISH)

πŸ“Š Training Data Sources

λ³Έ λͺ¨λΈμ€ μ•„λž˜μ˜ 곡개 데이터셋을 ν™œμš©ν•˜μ—¬ ν•™μŠ΅ 및 νŒŒμΈνŠœλ‹λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

1. Korean Message Dataset

  • Source: meal-bbang/Korean_message
  • Link: https://huggingface.co/datasets/meal-bbang/Korean_message
  • Description:
    ν•œκ΅­μ–΄ 문자 λ©”μ‹œμ§€(SMS)λ₯Ό μ€‘μ‹¬μœΌλ‘œ κ΅¬μ„±λœ λ°μ΄ν„°μ…‹μœΌλ‘œ,
    정상 λ©”μ‹œμ§€μ™€ 슀팸/사기성 λ©”μ‹œμ§€λ₯Ό ν¬ν•¨ν•˜κ³  μžˆμŠ΅λ‹ˆλ‹€.
    λ³Έ λͺ¨λΈμ—μ„œλŠ” 초기 λ‹¨κ³„μ˜ μ–Έμ–΄ 적응 및 μŠ€λ―Έμ‹± ν‘œν˜„ ν•™μŠ΅μ— ν™œμš©λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

2. KOR Phishing Detect Dataset

  • Source: Ez-Sy01/KOR_phishing_Detect-Dataset
  • Link: https://github.com/Ez-Sy01/KOR_phishing_Detect-Dataset
  • Description:
    μ‹€μ œ ν•œκ΅­μ–΄ μŠ€λ―Έμ‹± 및 ν”Όμ‹± 사둀λ₯Ό 기반으둜 κ΅¬μΆ•λœ λ°μ΄ν„°μ…‹μœΌλ‘œ,
    μŠ€λ―Έμ‹± 탐지 μ„±λŠ₯ ν–₯상을 μœ„ν•œ 핡심 νŒŒμΈνŠœλ‹ λ°μ΄ν„°λ‘œ ν™œμš©λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

⚠️ Data Usage Note

  • λ³Έ λͺ¨λΈμ€ 곡개적으둜 제곡된 λ°μ΄ν„°μ…‹λ§Œμ„ μ‚¬μš©ν•˜μ—¬ ν•™μŠ΅λ˜μ—ˆμŠ΅λ‹ˆλ‹€.
  • 데이터셋에 ν¬ν•¨λœ 개인 식별 정보(PII)λŠ” λͺ¨λΈ ν•™μŠ΅ κ³Όμ •μ—μ„œ μ§μ ‘μ μœΌλ‘œ μ‚¬μš©λ˜μ§€ μ•ŠμœΌλ©°,
    μ „μ²˜λ¦¬ 및 μΌλ°˜ν™” 과정을 톡해 νŠΉμ • κ°œμΈμ΄λ‚˜ 사둀λ₯Ό μž¬μ‹λ³„ν•  수 없도둝 μ„€κ³„λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

πŸ“ˆ Evaluation Results

λ™μΌν•œ ν…ŒμŠ€νŠΈ μ…‹ κΈ°μ€€μ—μ„œ μ•„λž˜ μ„±λŠ₯을 ν™•μΈν–ˆμŠ΅λ‹ˆλ‹€.

Model-only Evaluation (threshold = 0.5)

Metric Score
Accuracy 0.999
Precision (PHISH) 1.00
Recall (PHISH) 0.95
F1-score (PHISH) 0.97
  • **False Positive(μ˜€νƒ)**λ₯Ό μ΅œμ†Œν™”ν•˜λŠ” 데 쀑점을 λ‘” 섀계
  • μ‹€μ œ 운영 ν™˜κ²½μ—μ„œ λΆˆν•„μš”ν•œ 차단을 μ€„μ΄λŠ” 것을 λͺ©ν‘œλ‘œ 함

πŸš€ Recommended Usage

κΈ°λ³Έ μ‚¬μš© μ˜ˆμ‹œ

from transformers import pipeline

classifier = pipeline(
    "text-classification",
    model="donghyun95/kor-smishing-xlmroberta",
    truncation=True
)

classifier("λ³΄μ•ˆ κ°•ν™”λ₯Ό μœ„ν•΄ μ•„λž˜ 링크에 접속해 인증번호λ₯Ό μž…λ ₯ν•˜μ„Έμš”.")

πŸ› οΈ 운영 ν™˜κ²½ ꢌμž₯ μ „λž΅

  • Model-only score threshold: 70
    (pipeline 좜λ ₯ score Γ— 100 κΈ°μ€€)

  • λͺ¨λΈ μ μˆ˜κ°€ μ• λ§€ν•œ 경우 (예: 40 ~ 70 ꡬ간)μ—λ§Œ
    μ•„λž˜μ™€ 같은 μ΅œμ†Œν•œμ˜ 룰을 보쑰적으둜 μ μš©ν•˜λŠ” ꡬ쑰λ₯Ό ꢌμž₯ν•©λ‹ˆλ‹€.

    • 단좕 URL 포함 μ—¬λΆ€
    • OTP μž…λ ₯ + μ„€μΉ˜/κΆŒν•œ μš”μ²­
    • μ›κ²©μ œμ–΄ μ•± μ„€μΉ˜ μœ λ„

이 방식은 λ‹€μŒκ³Ό 같은 μž₯점을 κ°€μ§‘λ‹ˆλ‹€.

  • 정밀도(Precision)λ₯Ό μœ μ§€ν•˜λ©΄μ„œ
  • 운영 쀑 False Negativeλ₯Ό μ μ§„μ μœΌλ‘œ 보완 κ°€λŠ₯

⚠️ Limitations

  • μ‹ κ·œ μŠ€λ―Έμ‹± 문ꡬ λ˜λŠ” μƒˆλ‘œμš΄ μ‚¬νšŒκ³΅ν•™ νŒ¨ν„΄μ— λŒ€ν•΄μ„œλŠ”
    μ„±λŠ₯ μ €ν•˜κ°€ λ°œμƒν•  수 μžˆμŠ΅λ‹ˆλ‹€.

  • λ³Έ λͺ¨λΈμ€ λ³΄μ•ˆ νŒλ‹¨ 보쑰 도ꡬ이며,
    μžλ™ μ°¨λ‹¨λ³΄λ‹€λŠ” 경고·주의 μ•ˆλ‚΄ μš©λ„λ‘œμ˜ μ‚¬μš©μ„ ꢌμž₯ν•©λ‹ˆλ‹€.


πŸ›‘οΈ Ethical Considerations

  • λ³Έ λͺ¨λΈμ€ 사기 탐지 및 μ‚¬μš©μž 보호λ₯Ό λͺ©μ μœΌλ‘œ μ„€κ³„λ˜μ—ˆμŠ΅λ‹ˆλ‹€.
  • κ°μ‹œ, κ²€μ—΄, λΆ€λ‹Ήν•œ μžλ™ 차단을 μœ„ν•œ μ‚¬μš©μ„ μ˜λ„ν•˜μ§€ μ•ŠμŠ΅λ‹ˆλ‹€.
  • μ˜€νƒ(False Positive) κ°€λŠ₯성을 κ³ λ €ν•˜μ—¬
    μ‚¬μš©μž 확인 μ ˆμ°¨μ™€ ν•¨κ»˜ μ‚¬μš©ν•˜λŠ” 것이 λ°”λžŒμ§ν•©λ‹ˆλ‹€.

πŸ“š Citation

연ꡬ λ˜λŠ” ν”„λ‘œμ νŠΈμ—μ„œ λ³Έ λͺ¨λΈμ„ ν™œμš©ν•˜μ‹€ 경우,
μ•„λž˜μ™€ 같이 μΈμš©ν•΄ μ£Όμ„Έμš”.

@misc{donghyun95_kor_smishing_xlmroberta,
  author = {Donghyun},
  title = {Korean Smishing Detection Model based on XLM-RoBERTa},
  year = {2026},
  url = {https://huggingface.co/donghyun95/kor-smishing-xlmroberta}
}

πŸ‘€ Author Donghyun

Hugging Face: https://huggingface.co/donghyun95

Downloads last month
10
Safetensors
Model size
0.3B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support