LLM Research

I’m currently seeking HiWi / research assistant opportunities related to large language models (LLMs),
especially in efficient fine-tuning, multimodal fusion, and agentic reasoning systems.

Technically, I have hands-on experience with
agent fine-tuning, MCP (Model Context Protocol), and parameter-efficient adaptation methods like LoRA, adapters, and quantized distillation.
My work often combines system-level engineering with research-driven experimentation
I like building things that help us understand how models think.

I’m particularly interested in:

  • Multimodal understanding — bridging language, vision, and structured data to ground reasoning.
  • Efficient fine-tuning — making large models smaller, faster, and adaptive without losing reasoning depth.
  • Interpretability — uncovering how internal activations, neurons, or adapters encode concepts.

Below are several recent projects that reflect my approach: integrating practical engineering with research curiosity.


Resource-Efficient Distillation of Qwen Models

📅 Jul – Sep 2025 · GitHub · Report PDF
Keywords: LLM Compression · Quantization · Knowledge Distillation · LoRA

Reproduced and extended a teacher–student distillation pipeline for Qwen2.5–0.5B-Instruct.
Combined LoRA fine-tuning with 8-bit quantization to compress the model from 3B to 0.5B parameters.
Achieved 75% of teacher accuracy with 6× lower compute — demonstrating scalable efficiency for mid-size LLMs.

Distillation Pipeline Illustration

K-Adapter Reproduction & Ablation Study

📅 Jul – Sep 2025 · GitHub · Report PDF
Keywords: Adapter Tuning · Knowledge Injection · PEFT

Re-implemented K-Adapter (ACL 2021) for knowledge injection via frozen PLMs.
Ran controlled ablations on adapter layer depth, placement, and task overlap.
Found mid-layer adapters offered optimal trade-offs between factual recall and model stability.


Hierarchical Character-Level Language Model

📅 Jul – Sep 2025 · GitHub · Report PDF
Keywords: Hierarchical LM · Cache Mechanisms · Representation Learning

Re-implemented HCLM+Cache (Kawakami et al., 2017) in PyTorch to study word reuse in open-vocabulary settings.
Introduced vectorized computation and continuous cache management, improving throughput 3.8× and reducing validation BPC by 11.8%.
Showed that the cache component contributes most to long-range linguistic coherence.


LoRA-Driven Anime Style Generation

📅 Jul – Sep 2025 · GitHub · Report PDF
Keywords: LoRA · Diffusion Models · Multimodal Adaptation

Benchmarked LoRA against Textual Inversion and DreamBooth for anime-style generation using Stable Diffusion v1.5.
Showed that LoRA achieved ~47 FID reduction under small (100-image) datasets while using <2% trainable parameters.
Explored style blending by interpolating LoRA checkpoints, demonstrating compositional flexibility.


Plant Recognition with CNNs and Transfer Learning

📅 Jul 2025 · Report PDF
Keywords: Computer Vision · CNN · Transfer Learning

Developed and compared ResNet, EfficientNet, and Vision Transformer baselines on a plant classification dataset.
Demonstrated how transfer learning significantly improves data efficiency —
serving as an early exploration into model generalization and feature reuse.


Self-Conditioned Generation (Reimplementation)

📅 Jan – Mar 2025 · GitHub · Report PDF
Keywords: Mechanistic Interpretability · Hidden-State Control

Reproduced Self-Conditioned Pretrained LMs (ICML 2022) focusing on internal feedback loops in generation .
Analyzed how hidden-state reuse improves fluency and stability over standard autoregressive decoding.
Demonstrated controllable text steering without external conditioning or retraining.


Data Contamination in Large Models

📅 Feb 2025 · Report PDF
Keywords: LLM Evaluation · Dataset Integrity · Temporal Robustness

Investigated contamination within open LLM evaluation datasets.
Implemented prefix-based detection and cross-version comparison pipelines to quantify leakage effects.
Provided empirical support for stricter dataset curation in benchmark design.


MLLM: Towards Multimodal Language Models

📅 Mar 2025 · Report PDF
Keywords: Multimodal LLMs · Vision–Language Alignment

Surveyed and analyzed recent multimodal LLM architectures (e.g., BLIP-2, Flamingo, LLaVA).
Focused on how frozen-language backbones interact with visual Q-formers and alignment objectives.
Discussed open challenges in cross-modal grounding and scalability of visual–text fusion.


Temporal Reasoning in Clinical NLP

📅 Mar 2025 · Report PDF
Keywords: Temporal Reasoning · Clinical NLP · Knowledge Graphs

Reviewed how LLMs handle temporal information in clinical narratives.
Summarized key challenges in chronological inference and discussed methods like TIMER-Instruct and temporal knowledge graphs for improving event sequencing and model reliability.


✨ Research Focus

Across all my work, I’m drawn to one big idea:
how to make large models more adaptive, explainable, and grounded.

Whether it’s through agent fine-tuning, multimodal fusion, or parameter-efficient learning,
I enjoy working at the intersection of research and real systems
turning theoretical questions into reproducible, working prototypes.

If your lab or team is exploring similar topics, I’d love to contribute as a research assistant or HiWi
and help bridge engineering practicality with scientific insight.