Project Awards

Congratulations to the following teams, who produced exceptional, award-winning projects!

Best Default Project

Best Custom Project

Outstanding Default Projects

Outstanding Custom Projects

Custom Projects

Project nameAuthors
20 Questions for Code: Improving Code Generation with Information-Theoretic ClarificationAlexandra Suriya Kim, Julia Xi, Ria Garg
A Bigger Catch: Fine-Grained Curriculum Standards Alignment on the MathFish BenchmarkMayank Sharma, Teah Shi, Xinman Liu
Adapting Language Models for Low-Resource GPU Kernel ProgrammingAnnmaria Antony, Laasya Konidala, Natalia Pahlavan
Adaptive Test-Time Compute for Efficient Reasoning in Language ModelsRyan Tan
Adaptive Test-Time Compute for Pedagogically Grounded Reasoning in LLMsIsha Jain, Jason Sejin Chon, Medhya Goel
Agents Don’t Always Do What They ThinkMark William Gernitis
Always-On Learning Companion: Proactive Multimodal Tutoring for Everyday Study ScenariosChenyue Li, Haowen Wang, Zhen Jia
Analyzing Robustness and Context Use in Clinical Natural Language InferenceRyan Minh-Tri Le
Attention Modifications for Improved AdaptationJerry Yin, Michael Jang
Auditing Model-Generated Privacy Benchmarks: Do Synthetic Evaluations Reflect Real User Privacy Norms?Selena She
Benchmarking and Improving Generative Diversity in Language Models via Diverse Preference OptimizationAnnika Kaul Singh, Shyam Sai Bethina
Betting on Reasoning: Predicting Forecast Reliability in Prediction MarketsRahul Rejeev
Beyond Bradley-Terry: Random Logit Preference Modeling for RLHFJunyi Liu
Beyond Knowledge: Syntactic Complexity as a Bottleneck for Reasoning in "Bracket City" PuzzlesAmrita Malhotra
Bootstrapping Reasoning in Compact Language Models: A Multi-Stage Reinforcement Learning Pipeline with Targeted Failure RepairJoseph Li, Max Luis Rodriguez, Victor Chen
Bootstrapping Safety-Aligned Reasoning in Small Language Models via Self-InstructJ Yim, Komal Vij, Tim Jing
Building A Contextual Reasoning Aware Social-Intel Agent with Reinforcement LearningBinbin Li, Da Sun, Ying Lu
Burst: Multi-Agent System for High-Quality Temporal Content GenerationJeffrey Hao Wang
Can Coding Agents Manage Their Own Memory?Jerry Wang, Ryan Wang, Sameer Agrawal
Cartoon Caption Humor Quality Assessment and Generation with DPO and LoRAIsaiah Flores, Katherine Ha Wang
Causal Transfer of Semantic Operators Across Transformer Language ModelsShivatmica Murgai
Childproofing LLMs with Contrastive Activation AdditionRosemary Mingrui Jiang
Childproofing LLMs: A Comparative Analysis of CAA, ReFT, and DPO for Safety AlignmentAlice Zhu, Anya Han Zhang
CoFi-PG: Counterfactual Policy Gradients under Filtered Feedback for Multi-Agent LLMsElai Ben-Gal, Stela Tong
Cognitive Compression: Hierarchical Chain-of-Thought for Efficient LLM ReasoningAnuj Jamwal
Collaborative Dynamic Cheatsheet: Multi-Agent Test-Time Learning with Small Language ModelsErica Wang, Malvyn Lai
Compiler-in-the-Loop: Decomposing the Value of Static Verification for Low-Resource Code GenerationHlumelo Notshe, Joshua Martinez
Compositional Tool-Sequence Generation in Small Language ModelsBenji Warburton, Maanit Goel
Concept Training for Human-Aligned Language ModelsChristine Zhang
Confusion-Set Guided Retrieval for LLM-Constrained Brain-to-Text DecodingAndrew Su, Hyungjae Kim, Vincent Jinpeng Yip
Context Under Pressure: How Language Model Agents Should Save and Read Information Over Long InteractionsAlexander Owen Worley
Continuous Utility Direct Preference OptimizationMuhammad Ahmed Mohsin
Cost-Aware Escalation from Scalar Reward Models to Generative ModelsCole Yarbrough, Landon Renjiro Maka'ike Choy, Rui Chen
Curriculum-Based Fine-Tuning for Summarization of Endometriosis DataAli Hicham Tout
Data-Centric Control of Verbosity for DPO-Based Instruction AlignmentSusan Lee, Will Richard Alex Furlow
DeepRoot: Graph-Coordinated Multi-Agent ReasoningSean J Wang, Sijbren Manuel Kramer, Zijian (Carl) Ma
Designing a Conservative Humor Filter Can a Model Tell If an Image Caption Is Funny?Michael Roger
Diagnosing the Reversal Curse via Mechanistic Probing and Symmetric TrainingDeepti Gupta, Ke Huang, Rafael Cardoso Ferreira
Diversity-Incentivized GRPO for Constrained Arithmetic ReasoningGaurav Tyagi
Do Language Models implement compositional solutions for natural language understandingAhmad Jabbar
Do Long Contexts Help Legal Knowledge? A Case Study on US–China Securities RegulationYufei Peng
Does Fine-Tuning Hurt Cross-Platform Generalization in Depression Detection?Yanav Lall
Does Pedagogy Hurt Truth? Evaluating Educational Rewriting in MedicineCally Lin, Sasa Simic
Don’t Think About It: Activation Steering as Silent Defense Against Prompt InjectionGaurav Anand
Dynamic Ledger: Retrieval-Augmented Structured Memory for Test-Time LearningJerry Gu, Sabrina Yen-Ko, Shurui Liu
Dynamic Token Merging for Efficient Subword Encoder-Decoder TransformersChris Gu, Marco Andono Sie, Nathan Zhou
Dynamic Token Merging for Encoder-Only Transformers: Adapting MrT5’s Delete Gate to BERT and XLM-RoBERTaAronima Dass, Hiva Zaad, Tianhui Huang
Effect of Text Embedding Scale on GraphRAG AccuracyJon Valur Bjornsson
Emotional Arc Preservation in LLM Literary TranslationChloe Di Murdoch, Esidore Fajardo Eneinyang, Julia E Rhee
End-to-End Driving Trajectory Prediction with Vision-Language-Action ModelAnze Liu
Energy-Accuracy Trade-offs in Transformer-Based NLP Models: A Unified Benchmarking StudyThibaud Xavier Clement
Entropy-Triggered RAG: Optimizing Retrieval Efficiency via Token-Level Shannon EntropyYucheng Yao
Evaluating Eliminative Reasoning in LLM-Based Differential DiagnosisSeyun Bang, Tatiana Zhang
Evaluating JEPA for Natural Language TasksHenry Jingsong Zhou, Oleh Ivankiv, Yousef Hassan Ramadan
Evaluating Robustness of Large Language Models to AlgospeakHnin Yupar Mon, Thet Htar Thin Zar
Evaluating Robustness of Social Bias Detection to Lexical and LLM-Driven PerturbationsNomin-Erdene Bayarsaikhan
Evaluating User-Style Adaptation for Professional Text GenerationAndrea Ji Woo Nam Song
FACT: Attention Consistency Training Mitigates Sycophancy and JailbreaksEmma Sampietro, Justin Nicolas Hartenstein
FALCON: Factual-Aware Logical Consistency for Large Language Model Outputs via NLI-Guided Mixed Integer OptimizationRehan Raza Azam
Fast Compression versus Exact Recall: Investigating the Trade-offs Between Models in Specialized Reasoning TasksJerry Xiao, Nick Yan
Fast Vocabulary Transfer for African Languages in Multilingual Machine TranslationKailash Chandran Elumalai, Biya Brook
From Premise to Punchline: A Fine-tuned Model and “Writers’ Room” Framework for Saturday Night Live Sketch Script GenerationHannah Yu, Natalie Hampton
From Private Memory to Collective Intelligence: Collaborative Test-Time LearningJiaming Shen, Jiaxin Fang, Xinrui Jiang
From Symptoms to Syndromes: Development and Validation of Fine-Tuning Transformer Architectures for Genetic Neurological Disease DiagnosesAnushka Rawat, Ximing Gao, Yi Li
Generative Dialogue State Tracking with GPT-2 for Task-Oriented Service ConversationsVenu Madhav Samprathi Ram Prasad
Gold-Guided Programmatic Distillation for FinancialElana N Chen, Erica Zhao, Yun Dong
Grounded Go Commentary Generation via Expert Engines and Structured TerminologyYudong Chen
HALLU-NLI: Revisiting Natural Language Inference (NLI) Hallucination Detection Methods forLLM-Generated BiographiesNathania Elizabeth Lim, Sally Lee, Sarah Dong
Hash Routed Delta Patches for Fast Knowledge Updates in Small LLMsArash Hamzehlou
Hidden Signals: Hallucination Prediction in Medical QACatherine M Zhang, Christina Ba, Ina Kathleen Chun
How does fine-tuning change internal representations in an audio transcription model?Brandon Liu, Jason Hu, Jenny Jin
Improving Lean4 Autoformalization via Cycle Consistency Fine-tuningArsen Shebzukhov
Improving Scientific Reasoning in Small Language Models via Process Preference Re-RankingArya Gupta, Marianne Feng Liu
Investigating the Impact of Persona-Based System Prompts During SFT of Code LLMsTushar Aggarwal
Language-Augmented Flow Matching Policies for Robust Out-of-Distribution Robot ManipulationJeff Liu, Lucas Sosnick
Language-Conditioned Objectives for Task-Agnostic Preference Learning and Controller UpdatesKyeong-Won Park
Learning a Discriminator for Conceptual Diversity in LM OutputsHangoo Kang, James Liu
Learning Efficient Tool Orchestration with Language ModelsOrhun Akengin
Learning from Critiques: A Geometric Framework for Response ImprovementHaozhan Gao
Learning When to Speak: Teaching LLMs Silence Through Specialized DPO and DistillationAllison Sara John, Anthony D Argyropoulos, Yubo Ruan
Location, Location, Generation: Fine-Tuning a VLM for Real Estate DescriptionsCarey Chang, Niko Terebuh Ustin
Measuring the Measure: Mechanistic Prompt Sensitivity for LLM-Based Populism CodingJiehan Liu
Mechanistic Deconvolution of Memory and Context in Quantum Language ModelsNathan Roll
MedDistill: Improving Clinical LLM Performance Through Natural Language Tabular InsightsJoshua Logan Shunk, Patrick Ruibin Li
MGA: Mixed Gated Attention for Efficient Long Context AttentionJen Ha, Bharat Kumar
Mixture-of-Steering Vectors (MoSV): Sparse Gating for Compositional Hallucination MitigationDaniel Winston Lee, Olufeolu Oluwapelumi Kolawole, Vedant Malolan Srinivas
MoSA: Mixture-of-Specialized-Agents for Cost-Efficient Long-Document Question AnsweringHaseeb Ismail, Mert Karabiyik, Shayaan Memon
Multi-Lane Retrieval-Augmented Generation for Pharmaceutical Regulatory Dossier WritingOmar Ingi Halldorsson
MuTaP: Multi-Task Mutation Predictor via LoRA-Adapted ESM-2Aya Aburous, Jad Bitar
NanoVQAEllen Xu
Non-Toxic Trash-Talking Fantasy FootballAndrew Dana Lawlor, Xander William Russell
On-Policy Context DistillationDarynne Lee, Shizhe He, Simon Pritchard
Perceptual-Aware Spatial Scene Synthesis (PASSS)Karan Singh Soin, Na Young Son
Pinpointing Latent Planning in Language Models with Lightweight Mechanistic MethodsHarshvardhan Singh, Nick Rui, Nicole Ma
PocketSheet: Enhancing Test-Time Learning using Efficient Memory Augmentation in Small Language ModelsPrabhjot Singh Rai, Sakthivel Sivaraman
Practical and Interpretable Unfair ToS Detection: Comparing Legal-Bert, Linear Lexical Models, and Editable treesBasel AlKanjo
Practical Design Decisions Can Matter More Than Training Algorithm Choice: A Study of LLM-Based Rust Bug RepairEthan Charles Morgan
Precision Under Pressure: Pushing the Boundaries of the Accuracy-Efficiency Frontier in Question Answering with Mixture-of-DepthsHaoyue Yang, Jan Miroslaw Kopanski, Soha Sultan
Preference-Based Alignment of Code Generation for MCP Server DevelopmentKristjan Dagur Egilsson, Rami Ratl Mrad
Progressive Screenplay Narrative Understanding via Contrastive LearningLuca Thomas Wheeler
Quantized Pre-training for Small Mixture-of-ExpertsRaghavendra Pranith Koppula
RAG-Based LLM Supported by Clinically Structured Re-Ranking, RL-Tuned Retrieval, and Agentic Workflow for ED Triage PredictionCharlotte Louise Kramer, Isha Arora, Nino Alex Triandafilidis
Rapping in Role: A Study of Persona Robustness in Large Language ModelsEunice Hyeyun Jung, Megan Ja
Recursive Self-Improvement for Continual Adaptation in CodeAaditya Vikram Nalawade, Chandra Suda, Ethan David Goodhart
Reward Design for Medical Safety: Reducing Sycophancy via Truth-Weighted RLHFJillian Chang, Juli Huang, Michael Kuang Min Li
Scaling Test-Time Compute to Improve Formal Reasoning in Lean via Compiler FeedbackAdam Joseph Banks, Alexander Huang
Self-Distillation for Discrete Flow Map ConsistencySuchir Agarwal
Self-Improving Diffusion Large Language Models via Asymmetric Self-GuidanceTianlang Chen
Small Models Think Big: Toward Effective Memory Distillation for Small Co-ScientistsJaanak Prashar, Renn Su, Summer Olivia Royal
Structural Line Markers and Multi-Pass Reranking for GPT-2 Sonnet GenerationAalaap S Hegde, Mudit Baid, Rakshit Kaushik
SUMMEHRY: LLMs for Generating Temporal Patient VignettesArlina Shen, Asmita Sood, Eashan Monga
Support-Aware Retrieval of Evidence Passages for Community NotesDorian Scott Gulley, Dyllan Han
SYMBRION: Symbol Context and Dream Ego Relations Across Lifelong Dream Series as a Tool for PsychoanalysisBobby Rohrkemper, Chia-Wei Cheng
Test Time Training for Sample-Efficient Practical Molecular OptimizationAaron Chee-Hung Lee, Ishvi Mathai
Test-Time Training on Binary Sub-ProblemsAndrew Sung, Darrow Robert Hartman, Leo Li
The Efficiency Threshold: Few-Shot Prompting vs. LoRAAbi Lopez, Daniel Joseph Grossman, Shreyas Chikkanayakanahalli Seshadri
The Feasibility of Token-Level Compute Allocation across Depth in Pretrained TransformersAnjali Sreenivas, Yuchen Li
The Rosetta Probe: Cross-Lingual Syntactic Transfer in Monolingual English BERTAnanya Niharika Navale
Towards Robust Natural-Language Proof VerificationSlim Barkallah
TRACE: Tool-augmented Reasoning via Atomic Cheatsheet EditingArnold Tianyi Yang, Kyleen Liao, Roshen Sanjay Nair
Understanding Mechanisms of Sycophancy in Multi-turn InteractionsCamila Blank
Understanding Value Embeddings in GPT-2 Training SpeedrunsArihan Varanasi, Markus Zhang
Verified Anchor Selection and Adaptive Curriculum for Dynamic Cheatsheet MemoryMengqian Chen
Verified On-Policy Self-DistillationJack Li, Sophia Yinfan Li
Verifier-Guided Reasoning for Cryptic Crossword Clue SolvingAarav Arora, Caleb Youngjae Whang Choe, Shamit R Surana
Visceral Judgment: LLM Refusal through Affective StateNicolas Kennedy
Vision-Language Model Router for RoboticsJadelynn Kim Dao, Milan Ganai, Satvik Sharma
Where Reasoning Branches: How Preference Pair Construction Shapes DPO for Mathematical ReasoningDuy Nguyen

Default Projects

Project nameAuthors
A Study of SFT-DPO Interaction and LoRA vs Full Fine-Tuning in Small Language ModelsChristy Yang, Yuming Feng
Accelerated DPO Fine-tuning GPT-2 with Constructed DataJessie Ou, Weixin Yu
Accelerating Attention for GPT-2 Using FLASHATTENTION, Longformer, and cosFORMERDiego Sierra, Thomas Sarda, Tom-Eliot Jullien
AdamW The Last LLM-Bender: The Legend of LoRAAri Barbella-Blaha, Kieran Javier Barrett
Adapting GPT-2 for Sentiment Analysis, Paraphrase Detection, and Sonnet GenerationFiona Han, Samih Shaheen Qureshi, William Charles Rose
Adapting GPT-2 Through Fine-Tuning Across NLP TasksRitu Patil
Adapting Pretrained GPT-2 via LoRA: How Much Fine-Tuning Do We Actually Need?Zengmingyu He, Zerong Chen
Adaptive Mixture-of-Heads: Routing Attention Heads in GPT-2 with Fixed and Dynamic SparsityDavid Stutz, Ryder Fried
An Investigation of GPT-2 Applications and Training Improvements, and Exploring Multi-Token Entity PredictionsBen Wengreen, Bhavya Ashish Shah, Jeffrey Meng
Applying Direct Preference Optimization to Improve GPT-2 Sonnet GenerationAadhav Prabu
Beyond Full Fine-tuning: Finding the Limits of GPT-2 Efficient AdaptationAlexander Huayi Zhong, Kaitlyn Angel Kwan, Songyu Han
Build GPT-2Lucia Losada, Nicole Cortes
Build GPT-2Yuchan Guo, Yushi Feng
Build GPT-2Pengyu Mo, Shirley Yu, Yixiao Zhang
Building GPT-2Suzannah Dalton Wistreich
Building GPT-2 and Perfecting Performance with Low-Rank AdaptationChenyu Song, Juntao Cheng, Mingyang Li
Building GPT-2 for Paraphrase Detection and Sonnet GenerationYifan Guo
Building GPT-2 with Finetuning OptimizationsEthan R Lee, Ethan Y Lu, Jingyu Zhang
Building GPT-2: Revisiting a Key Milestone of NLPAndy Tianqi Wang, Darren Chan, Derek Yan
Circuit-Aware Analysis of LoRA Fine-Tuning: What Changes, Where, and Why?Nathan Maidi
Cloze-Style Paraphrase Detection and Sonnet Generation with GPT-2: Exploring LoRA and Decoding StrategiesNick Fursa
Co-Adaptation in LoRA: Target Placement Effects and Inter-Module Interactions in GPT-2Shekhar Sharma
Comparing ReFT and LoRA on Classification and Generative Tasks with GPT-2Ryan Patrick Catullo
Cost–Performance Tradeoffs for GPT-2 Fine-Tuning: A Case Study on Paraphrase and Sonnet ContinuationRicardo Ruiz
CS 224N Default ProjectKayla Li, Yaojing Huang He
Cutting Out the Middleman: Direct Preference Optimization for Paraphrase Detection and Sonnet GenerationJustin Yuankai Leong, William Li
Data Efficient Fine-Tuning and Alignment of GPT-2Aryaman Gupta, Joseph Lee, Zeyuan Feng
Default Final Project: Efficient Adaptiation of GPT-2 via LoRAPedro Gaspar Pires
Direct Preference Optimization for Constrained Generation and Classification in GPT-2Jingxiong Zhao, Weining Li
Direct Preference Optimization for Improving Sonnet GenerationGio Ty
Direct Preference Optimization: From Paraphrase Detection to Sonnet GenerationFlorencio Paucar Sedano
Does the Optimizer Matter? LoRA vs Full Fine-Tuning in NLPAndy Dimnaku
DoRA the ExplorerCayden Gu, Imogen Lee
DoRA: Parameter-Efficient Fine-Tuning for GPT-2 on Cloze Paraphrase Detection and Sonnet GenerationAniket Gupta, Anjani Pangal, Mallika Parulekar
DPO for Structural Sonnet Generation and Paraphrase Detection with GPT-2Daniel Marcelo Mottesi, Diego Bustamante, Jason McLeod Amsler
Effects of Quantization on GPT-2 SmallIsabella Lynne Jordan
Efficiency and Inference: A Comparative Study of PEFT and Full Fine-TuningSanyam Gupta
Efficiency in GPT-2: Parameter Adaptation, Quantization, and Synthetic Data AugmentationAbhinav Chinta, Ethan Hersch, Ryan D'Cunha
Efficiency–Performance Trade-offs in LoRA-family: Fine-Tuning Methods for GPT-2Christine Li, Jason Yan, Justin Li
Efficient Adaptation and Structure-Aware Post-Training of GPT-2 for Paraphrase Detection and Sonnet GenerationBrandon Michael Kunitzer, Koa Lanakila Chang
Efficient Alignment Is All You NeedLingbo Duan, Shatong Zhu, Yufei Liu
Efficient Fine-Tuning and Alignment of GPT-2 for Downstream NLP TasksAdam Alhousiki, Kamal Mohammed ElMallah, Tommy Leong
Efficient Fine-tuning of GPT-2 for Paraphrase Detection and Sonnet GenerationJonathan You
Efficient Fine-Tuning of GPT-2 via Low-Rank Adaptation (LoRA)Min Zhang, Shang Gao, Shang Gao
Efficient Fine-Tuning of GPT-2: LoRA, Hyperparameter Search, and Scaling for Paraphrase Detection and Sonnet GenerationBrian Sha
Efficient Steering and Preference Alignment: Applying LoReFT and DPO to a Custom GPT-2 ArchitectureHaonan Zhu
Encoding Task Structure via Attention Biases and Adaptive ComputationDario Gaitzi Soatto
Enforcing Rigid Syntax: Using LoRA to Adapt GPT-2Monami Dutta Gupta
Enhanced Hybrid Search for LLM Hyperparamter OptimizationAaron Michael Sequeira, Avery Graham Voss, CJ Indart
Evaluating LoRA for Efficient GPT-2 Fine-TuningRaymond Ruimeng Llata, Vania Chow
Evaluating Low-Rank Adaptation and Nested Low-Rank Architectures for Paraphrase Detection and Sonnet GenerationIan Yue-Ran Chen
Evaluating Low-Rank Representation Finetuning for GPT-2 Downstream TasksAlvin Ayuyo
Evaluating Performance, Efficiency, and Memory Trade-offs in GPT-2 Attention MechanismsDevon Thomas Johnston Smith, Lily Annabelle Bailey
Exploring decoding and efficiency strategies for GPT-2Stephanie Stephanie Vezich Tamayo
Exploring LoRA Variants With GPT-2George Danchen Song, Justin Choo
Exploring Low-Rank Adaptation for Efficient GPT-2 Fine-TuningAndy Zhang, Yi Lu
Exploring Parameter-Efficient Fine-Tuning for Paraphrase Detection with GPT-2Krisha K Chokshi
Extending GPT-2 for Informal and Slang Aware Language UnderstandingDhruv Darshan Naik, Ruby Hernandez
Fairness-Aware Fine-Tuning of GPT-2 for Paraphrase DetectionDeonna Owens
Fine-tuning GPT-2 for Sentiment Analysis, ParaphraseLiliana Carolina Santos-Deonizio
Fine-Tuning GPT-2 for Sentiment Analysis, Paraphrase Detection, and Sonnet Generation with Parameter-Efficient AdaptationCarl Liu, Zikun Zhu
Fine-Tuning GPT-2 for Sentiment Analysis, Paraphrase Detection, Sonnet Generation and Political Affiliation DetectionAnna Wu, Iris Zixiao Xu, Samantha Malowane Leventis
Fine-Tuning GPT-2 for Sentiment, Paraphrase, and Sonnet TasksWalter Lopez Chavez
Fine-tuning GPT-2 with LoRAManish Agarwal, Pierce Cailean Sayer Mullin
Fine-tuning GPT-2 with LoRA and DPO for Accurate Classification and Constrained GenerationShaoxiong Zhang
Fine-Tuning GPT-2: A Playground for Discriminative and Generative Adaptation TasksAditi Somayajula, Sahithi Ankireddy
Fine-Tuning, Alignment, and Efficient Adaptation of GPT-2 for NLP Downstream TasksAhmed Mohamed Hassan Khidre Elsherbiny, Izhan Hamza, Patrick Wang
FlashAttention-Enhanced GPT-2 for Paraphrase Detection and Sonnet GenerationKatie Liu, Norah Asemota, William Yang
From Detection to Generation: Fine-tuning Large GPT2 Models for Paraphrasing and PoetryAnna Guo
From-Scratch GPT-2 and Efficient AdaptationBryan Alexis Pineda, Michael James Nixon
Full Fine-Tuning v. LoRA: Parameter-Efficient Adaption of GPT-2 for Paraphrase Detection and Sonnet GenerationMegha Bindiganavale, Rydham Goyal
GaLore: Gradient Low-Rank ProjectionChung-Suen Stephen Chan
GDPO for GPT-2Ahmed Sherif Ahmed Elbakry Mohamed
GPT-2 Default Project with Attention-only LoRA for Paraphrase DetectionYiqing Liu
GPT-2 Implementation and SpeedupAlexia Huang, Qi Wu
GPT-2 with LoRA OptimizationIllia Shkirko, Janhavi Purkar, Zhang Bai-han
GPT-2 with Varying Attention MechanismsAneesh Akella
GPT2 Optimization with PEFT and DPOKiran Sun
GRPOET-Rank: Group Relative Policy Optimization with External Text-RankingEric Liang, Jamin Jia-Ming Xie, William Z Liu
Hardware-Aware Self-Attention for GPT-2: A FlashAttention-based StudySiri Garudanagiri Virupaksha
Implementing a GPT-2 Decoder for Text Generation, Classification, and Paraphrase DetectionAlma Oralia Minerva Cooper, Antra Nakhasi, Louis Weisdorf
Implementing and Extending GPT-2 for Multi-Task NLP Applications: A Parameter-Efficient Fine-Tuning PerspectiveIsabel Li, Lianyu Yao, Yunjie Xu
Implementing and Fine-Tuning GPT2 for Sentiment Analysis, Paraphrase Detection, and Sonnet GenerationZhenghui Chen
Improving GPT-2 Fine-Tuning through Parameter-Efficient Adaptation and Preconditioned OptimizationAkhilesh Varadan Balasingam, Georgios Mikos
Improving GPT-2 Fine-Tuning with Direct Preference Optimization for Sonnet Generation and Paraphrase DetectionAnna Gutowska, Nicolas Bejar Arambula, Petru Cristian Budianu
Improving GPT-2 with Reinforcement Learning from AI Feedback: Automated Judges for Aligned Sonnet GenerationJason Meng, Shinnosuke Yagi
Improving Performance and Efficiency of GPT-2 on Sonnet Generation and Paraphrase Detection TasksDivya Bhojraj
Investigating Structure Aware Decoding and Cloze Style Classification for Robust GPT 2 Fine TuningAnya Von Diessl
KV Caching and Speculative Decoding at GPT-2 Scale: Acceptance, Cost Ratio, and the Limits of SpeedupAmulya Parthasarathy
Lather, Rise, Repeat: The Shampoo OptimizerRicky Javier Rios
Learning to Rhyme with Token-Weighted DPOFisher Marks
Leveraging GPT-2 for Multiple Downstream NLP Tasks: Classification and GenerationDavi Ferreira Veronese
Longformer-Style Sparse Attention for GPT-2Hoang D Nguyen, Peter Martin Alisky
LoRA vs LoReFT: Parameter-Efficient Fine-Tuning of GPT-2Allen Yuan, Andrew Wooyong Chung, Ryan Joonwon Suh
Lora-Enhanced GPT-2 with DPO for Sonnet GenerationFilip William Henriksson, Krish Maniar, Nicholas Simon Allen
Low-Rank Adaptation and Preference Optimization for Accessible Multi-Task GPT-2 Fine-TuningMahathi Mangipudi, Taylor Elizabeth Hamilton-Hankins, Tyler Kinh Ho
Low-Rank Adaptation for Efficient GPT-2 Fine-Tuning: Evaluation Across Classification and Generation TasksVivek Tiwari
Low-rank fine-tuning of the GPT-2 modelRongge Yan
Memory-Efficient Transformer Attention via Tiled FlashAttention-style ImplementationSagar Kapare
Metric-Aligned Sonnet Generation with LoRA and Self-Critical RL Fine-tuningTimothy Yu, Xiang Wan
MiniGPT: Implementation, Fine-Tuning, and Extensions for Constrained GenerationShiwei Que
Modernizing GPT-2: Integrating Low-Rank Adaptation, FlashAttention, and Multi-Token Prediction for Efficient Sonnet GenerationManan Sheth, Sanjay Dixit Bhuvanagiri
Optimizing GPT-2 for Downstream Tasks: An Exploration of PEFT, Preference Optimization, and SMARTAnastasiya Masalava, Eva Casto, Michael Rybalkin
Parameter-Efficient Adaptation of GPT-2 Across Classification and Generation TasksIsaias Martinez, Kristine Ma, Varsha Saravanan
Parameter-Efficient Adaptation of GPT-2 across Discriminative and Generative TasksJunran Jia, Xianya Fu
Parameter-Efficient Fine-Tuning for GPT-2: Comparing LoRA and ReFT on Paraphrase DetectionMatthias Jiro Walther, Ngoc Nguyen
Parameter-Efficient Fine-Tuning of GPT-2 for Classification and Text GenerationRyan He
Parameter-Efficient Fine-Tuning of GPT-2 Using DoRALinika Goel, Mindy Kay Harkness
Parameter-Efficient Fine-Tuning of GPT-2 using Low-Rank Adaptation (LoRA)Chris Alexander Perez
Parameter-Efficient Fine-Tuning of GPT-2 with LoRAChloe Yuri Jeon, Erick Angelo Ramirez
Parameter-Efficient Fine-Tuning of GPT-2 with LoRA: A Systematic Study of Rank, Scale, and Learning RateCuiyuanxiu Chen
Parameter-Efficient Fine-Tuning of GPT-2: Comparing LoRA, LoReFT, and Prefix Tuning Across Classification and Generation TasksSally Wang, Zijian Luo
Parameter-efficient fine-tunings for Downstream Adaptation of GPT-2Peiyu Li, Zefang Zhou
Parameter-efficient Finetuning and Preference Optimization of GPT-2 for Downstream TasksPankaj Rajak
Paraphrase Detection and Sonnet Generation using GPT-2Dongyu Jia, Omar Walid Ayoub
Paraphrase Detection and Sonnet Generation with LoRA and DPODiya Bhattacharjee, Jaagat Prashar, Kyle Tianshi
Preference Optimization for Parameter-Efficient Multi-Task Learning of GPT-2Jiecong Tan, Mark Yang
Preference-Optimized GPT-2 for Cloze-Style Paraphrase Detection and Sonnet Generation via DPO with GPT-Scored PairsAndrew Samuel Park, Arun J Moorthy, Welton T Wang
Preference-Tuning GPT-2 with LoRA and DPO for Classification and PoetryHiromichi Murakami, Yuliia Murakami
QLoRA Fine-Tuning for Reliable Structured Tool Calls with GPT-2Jay Khemchandani
Quadapter: Adapter for GPT-2 QuantizationEthan Cohen, Wesley Bian
Quantization of GPT-2 for Running on Edge DevicesGordy D Sun
Rank, Bits, and Data: Efficient GPT-2 Adaptation for Paraphrase Detection and Sonnet GenerationKerui Lu, Silin Du
Re-implementing GPT-2 for Classification, Paraphrase Detection, and Sonnet GenerationSaanvi Reddy Thummalapally, Shani Su
Robust Fine-Tuning of GPT-2 with SMART Regularization, LoRA, and Enhanced Decoding for Downstream NLP TasksJake Klosowski, Kevin Stephen, Alex E Wurm
Robust GPT-2 Fine-Tuning via LoRA and Smoothness-Inducing RegularizationPooya Nabavi
SADPOSS G: Structure-Aware Direct Preference Optimization for Shakespearean Sonnet GenerationMario Felix Sumali
Second-Order Optimization for GPT-2 Fine-Tuning: Exploring K-FAC for NLP Downstream TasksIsabella Kai He
Sentiment Classification using GPT-2 RepresentationsSahaj Saini
SMART Regularization and DPO for GPT-2 Sonnet GenerationMac Broido
SMART-GPT2: Adapting SMART-Style Regularization for Decoder-Only Fine-TuningAustin Ho
SMART-GPT2: Adversarial Regularization for GPT-2 Fine-TuningBahram Y Mohmand, Noah Sabbavarapu, Zihan Wang
SOAP and Sonnets: Improving the Optimization Efficiency of GPT-2Kenna Zeng
Source, Relay, and Suppressor Heads in a Poetry Generation CircuitKai Wen, Shaoyi Zhang
Sparse ReFTJacob Daniel Householder
SPLoRA: Sonnet Generation and Paraphrase Detection with LoRAJoseph Rabara Bailey, Maya Vendhan
Streaming Sonnets: Efficient Generation with KV Caching and QuantizationCodey Codey Sun, Michael Yang
Style Steering of GPT-2 Sonnet Generation with DPOClaudia Perez D'Arpino
Task-Dependent Effects of Parameter-Efficient Fine-Tuning: A GPT-2-Based StudyWeiwei Wu
Task-Driven Fine-Tuning and Efficient Attention for GPT-2Sixian Du, Susan Li, Yuzhou Bian
Task-Specific Fine-Tuning Strategies for Improving GPT-2 Across Classification and Generative NLP TasksAustin Chen, Cheney Sang, Harris Alan Lee
Task-Specific GPT-2 Adaptation: Structured LoRA for Paraphrase Detection and DPO for Sonnet GenerationPuyang Du, Xijia Liu
TaskRankFabio Ibanez, Peter Jason Benitez
Teaching LLMs to Forget Bad Data with Controlled UnlearningHanyu Yang
The LoRA(x)Gerwin Delsocora Mateo, Ryan Da
Title of your projectArun Brian Morris Chhetri, Ian Luka Lasic-Ellis, Marcus Batt Kushner
Uncertainty-Aware Self-Training for Paraphrase Detection + Learned Reranking for Sonnet GenerationAlex M Michael, Luis Marc Botin-Sanz de Sautuola, Xander Coulter Hnasko
Weight Decomposition Matters: DoRA vs. LoRA for Small GPT-2 Task AdaptationSvea Drekshagen
Where Does LoRA Actually Help? Probing Layer-Wise Adaptation in GPT-2 for Paraphrase Detection and Sonnet GenerationMona Anvarihosseinabad