ts /

Tasin

I build the things I've watched people quietly route around. Most recently that meant an ITSM platform — now handling 4,000+ real users with real enterprise auth — which I shipped, solo, at the college where I was also the one answering the tickets.

01. Selected Work

02. Research

Published · Research Paper · Authors: KM Khalid Saifullah

This study explores how multi-agent interaction enhances autonomous vehicle (AV) decision-making in dynamic traffic environments. While traditional AV models focus on individual autonomy, real-world traffic scenarios often require collective behavior through inter-agent communication and coordination. To investigate this, we developed a graph-based simulation environment that enables vehicle agents to exchange information and reroute in real time in response to road obstacles. Our findings demonstrate that communication and adaptive rerouting significantly reduce average wait times and improve travel efficiency. Furthermore, we introduce a lightweight memory mechanism—Object Memory Management (OMM)—which allows agents to retain knowledge of previously encountered obstacles. This feature proved critical in avoiding routing loops and redundant decisions. Together, these results highlight the potential of communication- and memory-enhanced agents in creating resilient, cooperative AV systems capable of navigating complex and unpredictable traffic networks.

Autonomous VehiclesMulti-Agent SystemsV2V CommunicationDynamic ReroutingTraffic Simulation

Published · National Conference on Undergraduate Research (NCUR) · Authors: KM Khalid Saifullah

Sentiment analysis plays a crucial role in understanding developer interactions, issue resolutions, and project dynamics within software engineering (SE). While traditional SE-specific sentiment analysis tools have made significant strides, they often fail to account for the nuanced and context-dependent language inherent to the domain. This study systematically evaluates the performance of bidirectional transformers, such as BERT, against generative pre-trained transformers, specifically GPT-4o-mini, in SE sentiment analysis. Using datasets from GitHub, Stack Overflow, and Jira, we benchmark the models' capabilities with fine-tuned and default configurations. The results reveal that fine-tuned GPT-4o-mini performs comparable to BERT and other bidirectional models on structured and balanced datasets like GitHub and Jira, achieving macro-averaged F1-scores of 0.93 and 0.98, respectively. However, on linguistically complex datasets with imbalanced sentiment distributions, such as Stack Overflow, the default GPT-4o-mini model exhibits superior generalization, achieving an accuracy of 85.3% compared to the fine-tuned model's 13.1%. These findings highlight the trade-offs between fine-tuning and leveraging pre-trained models for SE tasks.

Sentiment AnalysisSoftware EngineeringTransformer ModelsGPTBERTFine-tuningNLP

03. Experience

  • Lead Software Engineer & Architect

    AtlasDesk 2025 — Present
  • IT Specialist, Solutions Architect

    The College of Wooster 2025 — Present
  • Data Scientist Intern

    Schneider Electric 2024
  • Software Engineer Intern

    Jomee Jomaa Inc. 2023

04. Leadership

  • Google Developer Student Club Co-Lead, The College of Wooster, 2024–2025. Ran workshops and hackathons for the campus developer community.
  • Stanford TreeHacks 2024 Co-built GovAI — an LLM-powered tool for querying government policy in plain language.
  • Residential Assistant The College of Wooster, 2022–2024. Mentored 50+ residents.