fokan commited on
Commit
a05fc30
·
verified ·
1 Parent(s): 3f87cbd

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +11 -36
README.md CHANGED
@@ -4,59 +4,34 @@ tags:
4
  - knowledge-distillation
5
  - pytorch
6
  - transformers
7
- - student-model
8
  base_model: unknown
9
  ---
10
 
11
- # Distilled Student Model
12
 
13
- This is a student model created through knowledge distillation.
14
 
15
- ## Model Details
16
-
17
- - **Architecture**: StudentModel
18
- - **Hidden Size**: 1024
19
- - **Number of Layers**: 12
20
- - **Modalities**: ['text']
21
- - **Created**: 2025-08-26T13:45:10.851453
22
-
23
- ## Teacher Models
24
 
 
25
 
 
26
 
27
  ## Training Details
28
 
29
- - **Strategy**: weighted
 
30
  - **Training Steps**: 5000
31
  - **Learning Rate**: 0.001
32
 
33
  ## Usage
34
 
35
  ```python
36
- from transformers import AutoModel, AutoConfig
37
-
38
- # Load the model
39
- model = AutoModel.from_pretrained("path/to/model", trust_remote_code=True)
40
- config = AutoConfig.from_pretrained("path/to/model")
41
-
42
- # Use for inference or further training
43
- outputs = model(input_ids)
44
- ```
45
-
46
- ## Retraining
47
 
48
- This model can be used as a student model for incremental training:
49
-
50
- ```python
51
- # Load as existing student for further distillation
52
- existing_student = "path/to/this/model"
53
- # Add new teachers and continue training
54
  ```
55
 
56
- ## Files
57
 
58
- - `pytorch_model.safetensors`: Model weights
59
- - `config.json`: Model configuration
60
- - `model.py`: Custom model architecture
61
- - `training_history.json`: Complete training history
62
- - `README.md`: This file
 
4
  - knowledge-distillation
5
  - pytorch
6
  - transformers
 
7
  base_model: unknown
8
  ---
9
 
10
+ # fokan/train-modle2
11
 
12
+ This model was created using knowledge distillation from the following teacher model(s):
13
 
 
 
 
 
 
 
 
 
 
14
 
15
+ ## Model Description
16
 
17
+ A distilled model created using multi-modal knowledge distillation.
18
 
19
  ## Training Details
20
 
21
+ - **Teacher Models**:
22
+ - **Distillation Strategy**: weighted
23
  - **Training Steps**: 5000
24
  - **Learning Rate**: 0.001
25
 
26
  ## Usage
27
 
28
  ```python
29
+ from transformers import AutoModel, AutoTokenizer
 
 
 
 
 
 
 
 
 
 
30
 
31
+ model = AutoModel.from_pretrained("fokan/train-modle2")
32
+ tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")
 
 
 
 
33
  ```
34
 
35
+ ## Created with
36
 
37
+ This model was created using the Multi-Modal Knowledge Distillation platform.