YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

Inference of LightXML(可独立运行)

本目录已经改造成独立推理包:

未分区文本 -> 分区模型 -> LightXML 多分区模型 -> 法条 TopK 预测

无需依赖项目外部的 LightXML_B/srclabel_group0.npydata/dataset_for_ml/legal_provision

目录结构

  • model/
    • provision_predictor/
      • groupy_legal_provision.best.pt
      • groupy_legal_provision.best.json
      • label_map.json
      • label_group0.npy
    • partition_predictor/
      • best_model.pt
      • vocab.json
      • model_meta.json
    • MODEL_INFO.md
  • code/
    • run_inference_pipeline.py(兼容入口)
    • prepare_demo_samples.py(兼容入口)
    • pipeline/
      • run_inference_pipeline.py(主脚本)
      • prepare_demo_samples.py
    • partition/judgment_partition_infer/(本地分区推理代码)
    • provision/lightxml_model.py(本地法条模型结构代码)
  • data/
    • input/sample_raw_input_5.jsonl
    • intermediate/sample_zoned_output_5.json
  • results/
    • sample_predictions_5.json

输入格式(未分区)

data/input/sample_raw_input_5.jsonl 每行一个 JSON:

  • id: 样本 ID
  • text: 未分区案件全文
  • labels: 可选真实法条列表(用于命中分析)

示例:

{"id":"sample_1","text":"(完整案件文本)","labels":["法条A","法条B"]}

一键推理命令

在项目根目录执行:

& '.\.venv\Scripts\python.exe' -m pip install -r 'LightXML_B\Inference of LightXML\requirements.txt'
& '.\.venv\Scripts\python.exe' 'LightXML_B\Inference of LightXML\code\run_inference_pipeline.py' `
  --model-path 'LightXML_B\Inference of LightXML\model\provision_predictor\groupy_legal_provision.best.pt' `
  --label-map-path 'LightXML_B\Inference of LightXML\model\provision_predictor\label_map.json' `
  --group-y-path 'LightXML_B\Inference of LightXML\model\provision_predictor\label_group0.npy' `
  --partition-model-path 'LightXML_B\Inference of LightXML\model\partition_predictor\best_model.pt' `
  --partition-vocab-path 'LightXML_B\Inference of LightXML\model\partition_predictor\vocab.json' `
  --partition-anchor auto `
  --input-file 'LightXML_B\Inference of LightXML\data\input\sample_raw_input_5.jsonl' `
  --output-zoned 'LightXML_B\Inference of LightXML\data\intermediate\sample_zoned_output_5.json' `
  --output-pred 'LightXML_B\Inference of LightXML\results\sample_predictions_5.json' `
  --max-items 5 `
  --device cuda

仅 CPU 推理时将 --device cuda 改为 --device cpu

输出说明

  • data/intermediate/sample_zoned_output_5.json
    • 包含分区边界、锚点状态、6 个分区文本
    • 分区键:z1_headerz2_factsz3_issuesz5_resultz6_dissentz7_other
  • results/sample_predictions_5.json
    • 包含每条样本的 TopK 法条预测和分数
    • 若输入中有 labels,会输出 hits_top5

重新生成 5 条演示输入(可选)

& '.\.venv\Scripts\python.exe' 'LightXML_B\Inference of LightXML\code\prepare_demo_samples.py'
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support