nielsr HF Staff commited on
Commit
c4e1bc2
·
verified ·
1 Parent(s): 4039aeb

Add metadata and improve model card

Browse files

Hi! I'm Niels from the community science team at Hugging Face. I noticed this model card was missing some key information and metadata. This PR adds the relevant pipeline tag and license to the metadata, along with links to the research paper, project page, and GitHub repository. I've also included a sample usage section based on the official documentation to help users get started.

Files changed (1) hide show
  1. README.md +34 -4
README.md CHANGED
@@ -1,10 +1,40 @@
1
  ---
 
 
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
5
  ---
6
 
7
- This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
- - Code: https://github.com/Intellindust-AI-Lab/EdgeCrafter
9
- - Paper: https://arxiv.org/abs/2603.18739
10
- - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
3
+ pipeline_tag: image-segmentation
4
  tags:
5
  - model_hub_mixin
6
  - pytorch_model_hub_mixin
7
  ---
8
 
9
+ # EdgeCrafter: ECSeg-L
10
+
11
+ EdgeCrafter is a unified compact Vision Transformer (ViT) framework designed for efficient edge dense prediction. This specific model, **ECSeg-L**, is optimized for instance segmentation on resource-constrained devices. It is part of the work presented in [EdgeCrafter: Compact ViTs for Edge Dense Prediction via Task-Specialized Distillation](https://arxiv.org/abs/2603.18739).
12
+
13
+ - **Paper:** [EdgeCrafter: Compact ViTs for Edge Dense Prediction via Task-Specialized Distillation](https://arxiv.org/abs/2603.18739)
14
+ - **Repository:** [https://github.com/Intellindust-AI-Lab/EdgeCrafter](https://github.com/Intellindust-AI-Lab/EdgeCrafter)
15
+ - **Project Page:** [https://intellindust-ai-lab.github.io/projects/EdgeCrafter/](https://intellindust-ai-lab.github.io/projects/EdgeCrafter/)
16
+
17
+ ## Model Description
18
+ EdgeCrafter addresses the performance gap between compact ViTs and CNN-based architectures like YOLO on edge devices. By using task-specialized distillation and an edge-friendly encoder-decoder design, EdgeCrafter models achieve a strong accuracy-efficiency tradeoff. ECSeg-L provides a high-performance balance for instance segmentation tasks.
19
+
20
+ ## Usage
21
+ To use this model, please refer to the [official GitHub repository](https://github.com/Intellindust-AI-Lab/EdgeCrafter) for installation instructions. You can run inference using the following command:
22
+
23
+ ```bash
24
+ cd ecdetseg
25
+ # Run PyTorch inference
26
+ # Make sure to replace `path/to/your/image.jpg` with an actual image path and provide the path to the weights
27
+ python tools/inference/torch_inf.py -c configs/ecseg/ecseg_l.yml -r /path/to/ecseg_l.pth -i path/to/your/image.jpg
28
+ ```
29
+
30
+ For loading models directly via the Hugging Face Hub, check the [hf_models.ipynb](https://github.com/Intellindust-AI-Lab/EdgeCrafter/blob/main/hf_models.ipynb) notebook in the repository.
31
+
32
+ ## Citation
33
+ ```bibtex
34
+ @article{liu2026edgecrafter,
35
+ title={EdgeCrafter: Compact ViTs for Edge Dense Prediction via Task-Specialized Distillation},
36
+ author={Liu, Longfei and Hou, Yongjie and Li, Yang and Wang, Qirui and Sha, Youyang and Yu, Yongjun and Wang, Yinzhi and Ru, Peizhe and Yu, Xuanlong and Shen, Xi},
37
+ journal={arXiv},
38
+ year={2026}
39
+ }
40
+ ```