You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We test on label 1 (100 test cases) with temperature 1.0 and top_p 0.97. Preliminary results show the importance of Triton-specific training for compilation and correctness.
69
69
70
+
Here is the training progress plot at 100 steps:
71
+
72
+

70
73
71
74
## 📂 Code Structure
72
75
73
-
Here are two main components of the code:
74
-
1.`nano_r1_script.py`is originally from [nano-aha-moment](https://github.com/McGill-NLP/nano-aha-moment/blob/f6384878831796fc29f560016e3cd570d264b823/nano_r1_script.py)and modified for our project, `kernel-coder`
75
-
2.[`KernelBench`](https://github.com/insop/KernelBench)is forked from [this repo](https://github.com/ScalingIntelligence/KernelBench) and modified for our project
76
+
The codebase consists of two main components:
77
+
1.`nano_r1_script.py`- Modified from [nano-aha-moment](https://github.com/McGill-NLP/nano-aha-moment/blob/f6384878831796fc29f560016e3cd570d264b823/nano_r1_script.py) for our `kernel-coder` project
78
+
2.[`KernelBench`](https://github.com/insop/KernelBench)- Forked and modified from [ScalingIntelligence/KernelBench](https://github.com/ScalingIntelligence/KernelBench)
0 commit comments