Skip to content
View SparkJiao's full-sized avatar
🎯
Focusing
🎯
Focusing

Highlights

  • Pro

Block or report SparkJiao

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
SparkJiao/README.md

Hi there 👋

  • 🌱 I’m a PhD candidate at Nanyang Technological University and the Institute for Infocomm Research (I2R), A*STAR.
  • 🔭 I’m currently working on machine reasoning with LLMs.

Here are some projects that I lead or contribute to:

  • PFPO: The source code for our work: Preference Optimization for Reasoning with Pseudo Feedback (ICLR 2025).
  • ProcessDPO: The source code for our EMNLP work with outstanding paper award: Learning Planning-based Reasoning by Trajectories Collection and Process Reward Synthesizing.
  • LLMSanitize: Our developed tool for detection of data contamination in Large Langauge Models.
  • llama-pipeline-parallel: A prototype and tutorial for deepspeed pipeline parallel training on Llama.
  • UNK-VQA: Our collected dataset for benchmarking the abstention ability of multimodal large models.
  • PandaLLM: One of the earliest projects exploring continue pre-training of Llama on Chinese corpus.
  • LogicLLM: One self-supervised pre-training method to enrich the logic-relevant knowledge of large language models.

Here are some projects for machine reading comprehension and information retrieval in the era of encoder-only pre-trained LMs:

  • MERIt: A self-supervised contrastive learning objective for logical reasoning.
  • LARCH: Conversational image search.
  • REPT: Retrieval-based pre-training for machine reading comprehension.
  • STM: A Self-Training Method for Machine Reading Comprehension with Soft Evidence Extraction.

Fangkai's github stats Top Langs

Pinned Loading

  1. dandelionsllm/pandallm dandelionsllm/pandallm Public

    Panda项目是于2023年5月启动的开源海外中文大语言模型项目,致力于大模型时代探索整个技术栈,旨在推动中文自然语言处理领域的创新和合作。

    Python 1k 79

  2. dpo-trajectory-reasoning dpo-trajectory-reasoning Public

    [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".

    Python 67 2

  3. llama-pipeline-parallel llama-pipeline-parallel Public

    A prototype repo for hybrid training of pipeline parallel and distributed data parallel with comments on core code snippets. Feel free to copy code and launch discussions about the problems you hav…

    Python 54 2

  4. MERIt MERIt Public

    [Findings of ACL 2022] Meta-Path Guided Contrastive Learning for Logical Reasoning of Text

    Python 29 5

  5. LogicLLM LogicLLM Public

    [NAACL 2024] Source code for paper "LogicLLM: Exploring Self-supervised Logic-enhanced Training for Large Language Models".

    Python 7

  6. microsoft/unilm microsoft/unilm Public

    Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities

    Python 20.7k 2.6k