Hao Wang / 王 昊
Hi, my name is Hao Wang and I am a master student at Department of Computer Science and Communications Engineering, Waseda University. I am currently researching natural language processing in Kawahara Lab.
Research interests: natural language processing, multilingual, multimodal, machine translation.
Email: conan1024hao[at->@]akane.waseda.jp
Education
Master of Engineering, Department of Computer Science and Communications Engineering, Waseda University. (Apr. 2023 - Mar. 2025, Sup: Daisuke Kawahara)
Bachelor of Engineering, Department of Computer Science and Engineering, Waseda University. (Apr. 2019 - Mar. 2023, Sup: Daisuke Kawahara)
Experience
Trainee, RIKEN AIP. (Apr, 2023 - , Tokyo, Sup: Shuhei Kurita)
Research Intern, CyberAgent AI Lab. (Mar. 2023 - , Tokyo, Sup: Tetsuro Morimura)
Research Assistant, Waseda University. (Nov. 2022 - Mar. 2023, Tokyo)
Software Engineer Intern, CyberAgent, Inc. (Feb. 2023 - Feb. 2023, Tokyo)
Software Engineer Intern, LINE Corp. (Aug. 2022 - Oct. 2022, Tokyo) [blog]
Software Engineer Intern, Fixstars Corp. (May. 2021 - Jul. 2021, Tokyo)
Software Engineer Intern, Morpho, Inc. (Oct. 2020 - Dec.2020, Tokyo)
Data Science Intern, MC Digital, Inc. (Jul. 2021 - Jan. 2022, Tokyo)
Data Science Hackathon First Place, P&G Japan. (Sep. 2021, Tokyo)
Publication
International Conference
Hao Wang, Hirofumi Shimizu and Daisuke Kawahara. Kanbun-LM: Reading and Translating Classical Chinese in Japanese Methods by Language Models. To be appear in ACL2023 (Findings). [paper] [github] [demo]
Domestic Conference
堀尾海斗, 村田栄樹, 王昊, 井手竜也, 河原大輔, 山崎天, 新里顕大, 中町礼文, 李聖哲, 佐藤敏紀. 日本語におけるChain-of-Thoughtプロンプトの検証. 2023年度人工知能学会全国大会(第36回, JSAI2023). [github]
王昊, 清水博文, 河原大輔. 言語モデルを用いた漢文の返り点付与と書き下し文生成. 言語処理学会第29回年次大会(NLP2023). [paper] [poster]
王昊, 中町礼文, 佐藤敏紀. 日本語の大規模な基盤モデルに対するLoRAチューニング. 言語処理学会第29回年次大会(NLP2023). [paper] [slide]
Other Presentations
近藤瑞希, 王昊, 井手竜也, 伊藤俊太朗, Ritvik Choudhary, 栗原健太郎, 河原大輔. 日本語BigBirdの構築. 日本語言語資源の構築と利用性の向上(言語処理学会第29回年次大会 併設ワークショップ, JLR2023). [slide]
Others
Competitive Programming
AtCoder highest rating: 1513
Codeforces highest rating: 1937
TopCoder highest rating: 1292
Data Science Competition (Kaggle)
M5 Forecasting – Accuracy: Estimate the unit sales of Walmart retail goods. (271/5558 Top5% Silver Medal)
Halite by Two sigma: Collect the most halite during your match in space. (64/1139 Top10% Bronze Medal)
Google AI4Code – Understand Code in Python Notebooks. (28/1135 Top5% Silver Medal)
U.S. Patent Phrase to Phrase Matching. (128/1889 Top10% Bronze Medal)
Open Source Pre-trained Models