News

2026

MarchNew blog post: On the Connection of Chomsky Hierarchy of Languages vs Data Curriculum Learning in LLMs. Should we show a language model the Navier-Stokes equations before it learns to add fractions?
JanuaryOur paper Stop Taking Tokenizers for Granted: They Are Core Design Decisions in Large Language Models got accepted at EACL 2026.

2025

10 DecemberAttending NeurIPS 2025 in Vancouver. See you there!
AugustMoved to the Ablations/Scaling Laws of the Nova series models.
AugustALLaM's arabic specific evals, AraEval got accepted at EMNLP'25. [paper][leaderboard][datasets].
MayJoined Amazon AGI-F to work on Multilingual Capability of Nova series models.
MayMoved to SF bay area. Roaming around Sunnyvale.
AprilLeft the ALLaM team (😭) at NCAI, SDAIA to join the Nova team at Amazon AGI-F. The ALLaM team has since moved to the newly formed Humain, which has raised a $100B sovereign investment to disrupt GenAI in the Middle East.
MarchPresenting ALLaM at the ICLR 2025, See you in Singapore.
AprilZeroSumEval accepted as ACL'2025 as Demo. Project moved to FAIR.
MarchCheckout the fully automated agentic evaluation paper: ZeroSumEval: An Extensible Framework For Scaling LLM Evaluation with Inter-Model Competition
JanuaryCheckout the new HLE benchmark. [paper] [data]

2024

25 DecemberReceived MIT "Innovators Under 35" Award.
17 OctoberOne paper has been accepted at EMNLP 2024 on evaluation. See you in Miami.
13 OctoberOur paper on scaling evaluation (`scalable oversight`) for frontier-class models has just dropped. Feel free to reach out to me for details.
10 OctoberWe announced ALLaM during the keynote of the Global AI Summit as one of the main priorities for sovereign AI in the Kingdom of Saudi Arabia. We also released the model and evaluation details for the 34B pretraining from scratch model. Keynote [YouTube Link].
22 JulyWe released our technical paper on ALLaM. Feel free to reach out to me if you have any queries.
22 MayMy lab's alignment work was recently released. ALLaM was revealed at IBM Think Keynote [Youtube Link]. ALLaM is a nationwide LLM effort of Saudi Arabia. Paper coming soon ...
15 MayOur paper on xCodeEval has been accepted at ACL 2024. Unfortunately, I won't be traveling to Thailand :(.
16 MayTwo paper accepted at ACL'24. See you at Bangkok.

2023

6 AugJoined SDAIA as a Senior Resarch Scientist
7 MayThree paper accepted at ACL'23. See you at Toronto.
15 MarchCheck out our recent pre-print xCodeEval: A Large Scale Multilingual Multitask Benchmark for Code Understanding, Generation, Translation and Retrieval paper.

2022

20 DecCheck out our recent pre-print SPT: Semi-Parametric Prompt Tuning for Multitask Prompted Learning paper.
18 DecCheck out our recent pre-print BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting paper.
26 NovReturned from Amazon d2l summer internship.
3 NovCheck out our recent pre-print Crosslingual Generalization through Multitask Finetuning paper.
1-OctOur paper (What Language Model to Train if You Have One Million GPU Hours?) got accepted at EMNLP'22 findings.
05-JulJoined Amazon d2l team as a summer intern.
11-AprCheck out our new ACL'22 workshop paper What Language Model to Train if You Have One Million GPU Hours?.
7-FebCheck out our pre-print PromptSource paper.
2-FebCheck out my recent talk on T0++ paper.

2021

19-NovOur paper T0++ got accepted as a spotlight paper at ICLR'22
15-OctOur paper Multitask Prompted Training Enables Zero-Shot Task Generalization is online. Model (T0++) , Dataset (P3)
22-SeptCheck out my Recent talk on Finetuned Language Models Are Zero-Shot Learners at NTU-NLP lab.
25-AugMy internship work from Amazon got accepted as a short paper in EMNLP.
3-MayTwo paper UXLA (in main conference) and AugVic (in findings) got accepted in ACL-IJCNLP-2021 .

2020

6-NovGet back to the PhD study after completing my internship with Amazon Lex Team.
15-OctOur paper LNMap is accepted in EMNLP-2020.
03-AugI will be starting my internship at Amazon, with Lex Team.
28-AugPassed my PhD Qualifying Examination (QE).
28-AprPreprint of our new paper LNMap released in Arxiv.
24-AprPreprint of our new paper MultiMix released in Arxiv.
25-MarGave a talk on mBART (paper ) in NTU-NLP.
13-FebPresenting our Cross-lingual-NER paper on AAAI-2020.
4-JanGot AAAI-2020 travel scholarship. See you in New York.

2019

9-NovI will be a Teaching Assistant, for Graduate Deep Learning Course of NLP , NTU
11-NovA paper accepted on AAAI 2020.
15-JunA paper accepted on ACL 2019.
25-JanShows MT system to industry and government stakeholders.
15-JanJoins NTU-NLP lab as a PhD student.

Experience

 
 
 
 
 
April 27, 2025 – Present
San Francisco Bay Area

Applied Scientist

Amazon AGI

Currently:
- I work on Ablation/Scaling laws for Nova series model
- I worked on Multilingual Training (both pretraining and post-trianing) of Nova series model

My current reasearch interest is:
- Anatomy of Pretraining (see: ALLaM, BLOOM [1] [2] )
- Alignment of LLMs. (see: T0, BLOOMZ, SPT)
- Robust evaluation of frontier models. (see: xCodeEval , ChatGPTEval [1] [2] [3])
 
 
 
 
 
August 6, 2023 – April 17, 2023
Saudi Arabia

Senior Research Scientist

National Center for Artificial Intelligence (NCAI), SDAIA

Training and Aligning Large Language Model. During my time here at NCAI, SDAIA,
- I developed the core technology behind ALLaM.
- I was the Core maintainer of ALLaM.
- I lead the Training Team of ALLaM.
For the work of ALLaM, I received MIT TR35 (Innovator Under 35) award.
 
 
 
 
 
July 5, 2022 – October 10, 2022
Santa Clara, CA

Applied Scientist Intern

Amazon Development Center U.S., Inc.

SPT: Semi-Parametric Prompt Tuning for Multitask Prompted Learning.
 
 
 
 
 
July 1, 2021 – January 28, 2022
East Palo Alto, CA (WFH)

Applied Scientist Intern (Part Time)

Lex Team, Amazon Web Services, Inc.

Cross-lingual Transfer learning under low resource and low parameter scenario.
 
 
 
 
 
August 3, 2020 – November 6, 2020
East Palo Alto, CA (WFH)

Applied Scientist Intern

Lex Team, Amazon Web Services, Inc.

Doing research on Cross-lingual Lex Bot.
 
 
 
 
 
September 15, 2018 – January 6, 2019
Bangladesh

Software Engineering Intern

Aubichol IT Limited

Worked on an early startup where the key responsibility was to design the architecture of sports analytic and Translation System.
 
 
 
 
 
September 15, 2017 – August 27, 2018
Singapore

Research Assistant

Nanyang Technological University

Doing Research on Natural Language Processing.
Responsibilities include:
* Research on MT, NER and Adversarial Training.
* Maintain internal GPU servers
* Maintain group web-site

Recent Posts

I have been making hot takes and hypothesizing about so many things to my friends, and to some extent I have always been too lazy to …

Publications

Quickly discover relevant content by filtering publications

Reframing tokenization as a core modeling decision in LLMs rather than a preprocessing step, arguing for context-aware tokenizer and …

A family of multimodal reasoning and generation models from Amazon AGI.