Skip to content

Commit a3b82ed

Browse files
committed
Add Jiang report
1 parent 20c4a46 commit a3b82ed

File tree

1 file changed

+9
-0
lines changed

1 file changed

+9
-0
lines changed

lichess.bib

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -527,6 +527,15 @@ @article{jenner:2024:evidence-lookahead-chess-neural-network
527527
eprint = {2406.00877},
528528
}
529529

530+
@misc{jiang:2023:building-natural-language-chess-engine-pretraining-instruction-finetunine,
531+
title = {Building a Natural Language Chess Engine with Pretraining and Instruction Fine-Tuning},
532+
author = {Bowen Jiang},
533+
year = {2023},
534+
url = {https://web.stanford.edu/class/archive/cs/cs224n/cs224n.1234/final-reports/final-report-169466939.pdf},
535+
note = {Stanford CS224N Custom Project, Winter 2023 (https://web.stanford.edu/class/archive/cs/cs224n/cs224n.1234/project.html)},
536+
abstract = {Although pretrained large language models (LLMs) can generate convincing natural language about games like chess, they lack positional and contextual knowledge and as such are poor game-playing agents. In this project, I utilize language pretaining; instruction fine-tuning, an additional training regimen with chess-specific tasks presented in natural language; and chain-of-thought prompting, a natural language description of problem reasoning prepended to the answer of a problem, to improve the performance of LLMs at chess move generation (validity/legality and quality of moves). I show that fine-tuned GPT-2-XL, a 1.5B parameter LLM, performs favorably well at move generation compared to ChatGPT with few-shot learning; I also validate the additional benefits of chain-of-thought prompting compared to plain prompts in ChatGPT while highlighting tradeoffs between the quality of natural language and the quality of chess when more verbose prompts are used in the smaller GPT-2-XL.},
537+
}
538+
530539
@misc{kapla:2025:generalized-multi-linear-models-dimension-reduction-tensor-valued-predictors,
531540
title = {Generalized Multi-Linear Models for Sufficient Dimension Reduction on Tensor Valued Predictors},
532541
author = {Daniel Kapla and Efstathia Bura},

0 commit comments

Comments
 (0)