Skip to content

Commit c6f4d7c

Browse files
Update README
1 parent f00ff4b commit c6f4d7c

File tree

1 file changed

+1
-37
lines changed

1 file changed

+1
-37
lines changed

README.md

Lines changed: 1 addition & 37 deletions
Original file line numberDiff line numberDiff line change
@@ -85,9 +85,6 @@ docker exec -it rllm-container bash
8585
Our work is done as part of [Berkeley Sky Computing Lab](https://skycomputing.berkeley.edu/). The rLLM team is generously supported by grants from [Laude Institute](https://www.laude.org/), [AWS](https://aws.amazon.com/), [Hyperbolic](https://www.hyperbolic.ai/) and [Fireworks AI](https://fireworks.ai/). We pay special thanks to [Together AI](https://www.together.ai/) for the research partnership and compute support.
8686

8787
## Citation
88-
89-
Citing rLLM:
90-
9188
```bibtex
9289
@misc{rllm2025,
9390
title={rLLM: A Framework for Post-Training Language Agents},
@@ -99,37 +96,4 @@ Citing rLLM:
9996
}
10097
```
10198

102-
Citing our prior work DeepScaleR, DeepCoder, and DeepSWE.
103-
```bibtex
104-
@misc{deepswe2025,
105-
title={DeepSWE: Training a State-of-the-Art Coding Agent from Scratch by Scaling RL},
106-
author={Michael Luo and Naman Jain and Jaskirat Singh and Sijun Tan and Ameen Patel and Qingyang Wu and Alpay Ariyak and Colin Cai and Tarun Venkat and Shang Zhu and Ben Athiwaratkun and Manan Roongta and Ce Zhang and Li Erran Li and Raluca Ada Popa and Koushik Sen and Ion Stoica},
107-
howpublished={\url{https://pretty-radio-b75.notion.site/DeepSWE-Training-a-Fully-Open-sourced-State-of-the-Art-Coding-Agent-by-Scaling-RL-22281902c1468193aabbe9a8c59bbe33}},
108-
note={Notion Blog},
109-
year={2025}
110-
}
111-
```
112-
113-
Citing DeepCoder:
114-
```bibtex
115-
@misc{deepcoder2025,
116-
title={DeepCoder: A Fully Open-Source 14B Coder at O3-mini Level},
117-
author={Michael Luo and Sijun Tan and Roy Huang and Ameen Patel and Alpay Ariyak and Qingyang Wu and Xiaoxiang Shi and Rachel Xin and Colin Cai and Maurice Weber and Ce Zhang and Li Erran Li and Raluca Ada Popa and Ion Stoica},
118-
howpublished={\url{https://pretty-radio-b75.notion.site/DeepCoder-A-Fully-Open-Source-14B-Coder-at-O3-mini-Level-1cf81902c14680b3bee5eb349a512a51}},
119-
note={Notion Blog},
120-
year={2025}
121-
}
122-
```
123-
124-
Citing DeepScaleR:
125-
126-
```bibtex
127-
@misc{deepscaler2025,
128-
title={DeepScaleR: Surpassing O1-Preview with a 1.5B Model by Scaling RL},
129-
author={Michael Luo and Sijun Tan and Justin Wong and Xiaoxiang Shi and William Y. Tang and Manan Roongta and Colin Cai and Jeffrey Luo and Li Erran Li and Raluca Ada Popa and Ion Stoica},
130-
year={2025},
131-
howpublished={\url{https://pretty-radio-b75.notion.site/DeepScaleR-Surpassing-O1-Preview-with-a-1-5B-Model-by-Scaling-RL-19681902c1468005bed8ca303013a4e2}},
132-
note={Notion Blog}
133-
year={2025}
134-
}
135-
```
99+
You may also cite our prior work [DeepScaleR](https://scholar.googleusercontent.com/scholar.bib?q=info:PrmBADk39GwJ:scholar.google.com/&output=citation&scisdr=CgIJFx-xEMCQ6zOgcuI:AAZF9b8AAAAAaPCmauIfzg8Rm9ImNYDad0uPUK8&scisig=AAZF9b8AAAAAaPCmahXsNqb1jTQBw2iPfw2vm9g&scisf=4&ct=citation&cd=-1&hl=en&scfhb=1), [DeepCoder](https://scholar.googleusercontent.com/scholar.bib?q=info:xpZNEPI6opAJ:scholar.google.com/&output=citation&scisdr=CgIJFx-xEMCQ6zOgjM8:AAZF9b8AAAAAaPCmlM_hb3S0tzBSVrRYBZYDLWg&scisig=AAZF9b8AAAAAaPCmlG109SG8d8230AiDP4jMxlw&scisf=4&ct=citation&cd=-1&hl=en&scfhb=1), and [DeepSWE](https://scholar.googleusercontent.com/scholar.bib?q=info:J9rT3SnY_aMJ:scholar.google.com/&output=citation&scisdr=CgIJFx-xEMCQ6zOg3D4:AAZF9b8AAAAAaPCmxD7Nl0xA_AcAeydpcE1BXCo&scisig=AAZF9b8AAAAAaPCmxE2Spzf5lf-2Toys5xEpnuA&scisf=4&ct=citation&cd=-1&hl=en&scfhb=1).

0 commit comments

Comments
 (0)