I specialize in Large Language Models (LLMs) and Natural Language Processing (NLP), with a focus on advancing the state-of-the-art in downstream tasks such as semantic understanding, timeline summarization, and grammatical error correction. My work bridges cutting-edge research (4 ACL publications) and production-grade systems (ByteDance, Apple), making AI accessible and practical. Last but not least, my first name "Geonsik" is pronounced as "Gun-Shik" or /kΚn.Ιik/. You can also just call me "GS" ππ»ββοΈ
- Fine-tuning & Optimization: LoRA-based PEFT, 4-bit quantization, hyperparameter tuning with W&B
- Model Deployment: vLLM serving, LangChain pipelines, ChromaDB integration, AWS Bedrock inference
- Models: Mistral, Llama 2/3, FLAN-T5, GPT family, OpenAI GPT-OSS-20B
- Structured Outputs: Using instructor library with Pydantic models for reliable LLM responses
- Research: Incremental clustering algorithms using LLM-based pairwise classification
- Grammatical Error Correction (GEC): Sequence-to-sequence & sequence tagging approaches
- Timeline Summarization (TLS): Event detection, clustering, and narrative construction
- Semantic Understanding: Word Sense Disambiguation (WSD), Words-in-Context (WiC)
- Email Classification: Topic-based email classification with RAG-enriched semantic search
- Transfer Learning: Encoder-only vs. decoder-only architectures for semantic tasks
- Scalable Web Applications: Flask, Streamlit, Bootstrap, Docker containerization, LAMP stack
- Microservices Architecture: GEC system with separate API and web interface modules, email processing pipelines
- Model Serving: Production-grade deployment of transformer models and LLMs for real-time inference
- RAG Pipelines: Retrieval-Augmented Generation with vector embeddings for semantic search
- Novel approach leveraging LLMs for incremental event clustering and timeline construction from text streams. Outperformed SOTA on 4 TLS benchmarks.
- Tech Stack:
PyTorchvLLMLlama-2-13BLangChainChromaDB
- Comprehensive framework demonstrating encoder-only models outperform decoder-only LLMs on word meaning comprehension tasks.
- Tech Stack:
PyTorchHuggingFace TransformersLoRAPEFTWandB
- End-to-end email processing pipeline with Streamlit web UI for Gmail integration, intelligent topic classification using AWS Bedrock LLMs, and AI-generated email thread summaries. Features incremental processing, LLM-powered topic attribute generation, RAG-enriched classification with semantic search, and complete lifecycle management (create, view, delete projects).
- Tech Stack:
PythonStreamlitAWS BedrockOpenAI GPT-OSS-20BinstructorPydanticGmail APIFAISSAmazon Titan EmbeddingsChromaDBLangChain
LLM Agent Evaluation [Code]
- Research toolkit for analyzing LLM agent trajectories on software engineering tasks.
- Tech Stack:
JupyterPythonAgent Frameworks
Algorithm Practice [Code]
- Self-contained archive of LeetCode solutions demonstrating strong algorithmic foundations.
- Tech Stack:
PythonData StructuresAlgorithms
-
From Moments to Milestones: Incremental Timeline Summarization Leveraging Large Language Models
Qisheng Hu, Geonsik Moon, Hwee Tou Ng
ACL 2024 (Main Conference) | [Code|Paper] -
Are Decoder-Only Language Models Better than Encoder-Only Language Models in Understanding Word Meaning?
Muhammad Qorib, Geonsik Moon, Hwee Tou Ng
ACL 2024 (Findings) | [Code|Paper] -
ALLECS: A Lightweight Language Error Correction System
Muhammad Reza Qorib, Geonsik Moon, Hwee Tou Ng
EACL 2023 (System Demonstrations) | [Code|Paper] -
WAMP: Writing, Annotation, and Marking Platform
Geonsik Moon, Muhammad Reza Qorib, Daniel Dahlmeier, Hwee Tou Ng
IJCNLP-AACL 2023 (System Demonstrations) | [Code|Paper]
- π Website: gsmoon97.github.io
- πΌ LinkedIn: linkedin.com/in/gsmoon97
- π Google Scholar: si3AXV8AAAA
- π¬ ORCID: 0009-0001-5646-466X
- π§ Based in: New York, NY

