Skip to content

ICTMCG/Awesome-Machine-Generated-Text

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 

Repository files navigation

Awesome Machine Generated Text

TOC

Large Scale Pre-training for Language Generation

Note: #params > 1B only

  • OpenAI

    • (GPT-2) Language Models are Unsupervised Multitask Learners. [paper] [blog]

      Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei and Ilya Sutskever.

    • (GPT-3) Language Models are Few-Shot Learners. [paper]

      Show AuthorsTom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever and Dario Amodei.
    • WebGPT: Browser-assisted question-answering with human feedback. [paper] [blog]

      Reiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, Xu Jiang, Karl Cobbe, Tyna Eloundou, Gretchen Krueger, Kevin Button, Matthew Knight, Benjamin Chess and John Schulman.

    • (InstructGPT) Training language models to follow instructions with human feedback. [paper] [blog]

      Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike and Ryan Lowe.

    • ChatGPT: Optimizing Language Models for Dialogue. [blog]

    • (GPT-4) GPT-4 Technical Report. [paper] [blog] [blog]

      OpenAI.

  • DeepMind

    • (Gopher) Scaling Language Models: Methods, Analysis & Insights from Training Gopher. [paper] [blog]

      Show AuthorsJack W. Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young, Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer, Richard Powell, George van den Driessche, Lisa Anne Hendricks, Maribeth Rauh, Po-Sen Huang, Amelia Glaese, Johannes Welbl, Sumanth Dathathri, Saffron Huang, Jonathan Uesato, John Mellor, Irina Higgins, Antonia Creswell, Nat McAleese, Amy Wu, Erich Elsen, Siddhant Jayakumar, Elena Buchatskaya, David Budden, Esme Sutherland, Karen Simonyan, Michela Paganini, Laurent Sifre, Lena Martens, Xiang Lorraine Li, Adhiguna Kuncoro, Aida Nematzadeh, Elena Gribovskaya, Domenic Donato, Angeliki Lazaridou, Arthur Mensch, Jean-Baptiste Lespiau, Maria Tsimpoukelli, Nikolai Grigorev, Doug Fritz, Thibault Sottiaux, Mantas Pajarskas, Toby Pohlen, Zhitao Gong, Daniel Toyama, Cyprien de Masson d'Autume, Yujia Li, Tayfun Terzi, Vladimir Mikulik, Igor Babuschkin, Aidan Clark, Diego de Las Casas, Aurelia Guy, Chris Jones, James Bradbury, Matthew Johnson, Blake Hechtman, Laura Weidinger, Iason Gabriel, William Isaac, Ed Lockhart, Simon Osindero, Laura Rimell, Chris Dyer, Oriol Vinyals, Kareem Ayoub, Jeff Stanway, Lorrayne Bennett, Demis Hassabis, Koray Kavukcuoglu and Geoffrey Irving.
    • (RETRO) Improving language models by retrieving from trillions of tokens. [paper] [blog]

      Show AuthorsSebastian Borgeaud, Arthur Mensch, Jordan Hoffmann, Trevor Cai, Eliza Rutherford, Katie Millican, George van den Driessche, Jean-Baptiste Lespiau, Bogdan Damoc, Aidan Clark, Diego de Las Casas, Aurelia Guy, Jacob Menick, Roman Ring, Tom Hennigan, Saffron Huang, Loren Maggiore, Chris Jones, Albin Cassirer, Andy Brock, Michela Paganini, Geoffrey Irving, Oriol Vinyals, Simon Osindero, Karen Simonyan, Jack W. Rae, Erich Elsen and Laurent Sifre.
    • (Chinchilla) Training Compute-Optimal Large Language Models. [paper] [blog]

      Show AuthorsJordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, Tom Hennigan, Eric Noland, Katie Millican, George van den Driessche, Bogdan Damoc, Aurelia Guy, Simon Osindero, Karen Simonyan, Erich Elsen, Jack W. Rae, Oriol Vinyals and Laurent Sifre.
    • (Sparrow) Improving alignment of dialogue agents via targeted human judgements. [paper] [blog]

      Show AuthorsAmelia Glaese, Nat McAleese, Maja Trębacz, John Aslanides, Vlad Firoiu, Timo Ewalds, Maribeth Rauh, Laura Weidinger, Martin Chadwick, Phoebe Thacker, Lucy Campbell-Gillingham, Jonathan Uesato, Po-Sen Huang, Ramona Comanescu, Fan Yang, Abigail See, Sumanth Dathathri, Rory Greig, Charlie Chen, Doug Fritz, Jaume Sanchez Elias, Richard Green, Soňa Mokrá, Nicholas Fernando, Boxi Wu, Rachel Foley, Susannah Young, Iason Gabriel, William Isaac, John Mellor, Demis Hassabis, Koray Kavukcuoglu, Lisa Anne Hendricks and Geoffrey Irving.
  • Nvidia & Microsoft

    • Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism. [paper]

      Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper and Bryan Catanzaro.

    • Turing-NLG: A 17-billion-parameter language model by Microsoft. [blog]

    • (Megatron-Turing NLG) Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model. [paper] [blog] [blog]

      Show AuthorsShaden Smith, Mostofa Patwary, Brandon Norick, Patrick LeGresley, Samyam Rajbhandari, Jared Casper, Zhun Liu, Shrimai Prabhumoye, George Zerveas, Vijay Korthikanti, Elton Zhang, Rewon Child, Reza Yazdani Aminabadi, Julie Bernauer, Xia Song, Mohammad Shoeybi, Yuxiong He, Michael Houston, Saurabh Tiwary and Bryan Catanzaro.
  • Google

    • (Meena) Towards a Human-like Open-Domain Chatbot. [paper] [blog]

      Daniel Adiwardana, Minh-Thang Luong, David R. So, Jamie Hall, Noah Fiedel, Romal Thoppilan, Zi Yang, Apoorv Kulshreshtha, Gaurav Nemade, Yifeng Lu and Quoc V. Le.

    • (T5) Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. [paper] [blog]

      Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li and Peter J. Liu.

    • mT5: A massively multilingual pre-trained text-to-text transformer. [paper]

      Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua and Colin Raffel.

    • (FLAN) Finetuned Language Models Are Zero-Shot Learners. [paper] [blog]

      Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai and Quoc V. Le.

    • GLaM: Efficient Scaling of Language Models with Mixture-of-Experts. [paper] [slides] [blog]

      Show AuthorsNan Du, Yanping Huang, Andrew M. Dai, Simon Tong, Dmitry Lepikhin, Yuanzhong Xu, Maxim Krikun, Yanqi Zhou, Adams Wei Yu, Orhan Firat, Barret Zoph, Liam Fedus, Maarten Bosma, Zongwei Zhou, Tao Wang, Yu Emma Wang, Kellie Webster, Marie Pellat, Kevin Robinson, Kathleen Meier-Hellstern, Toju Duke, Lucas Dixon, Kun Zhang, Quoc V Le, Yonghui Wu, Zhifeng Chen and Claire Cui.
    • LaMDA: Language Models for Dialog Applications. [paper] [blog]

      Show AuthorsRomal Thoppilan, Daniel De Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, YaGuang Li, Hongrae Lee, Huaixiu Steven Zheng, Amin Ghafouri, Marcelo Menegali, Yanping Huang, Maxim Krikun, Dmitry Lepikhin, James Qin, Dehao Chen, Yuanzhong Xu, Zhifeng Chen, Adam Roberts, Maarten Bosma, Vincent Zhao, Yanqi Zhou, Chung-Ching Chang, Igor Krivokon, Will Rusch, Marc Pickett, Pranesh Srinivasan, Laichee Man, Kathleen Meier-Hellstern, Meredith Ringel Morris, Tulsee Doshi, Renelito Delos Santos, Toju Duke, Johnny Soraker, Ben Zevenbergen, Vinodkumar Prabhakaran, Mark Diaz, Ben Hutchinson, Kristen Olson, Alejandra Molina, Erin Hoffman-John, Josh Lee, Lora Aroyo, Ravi Rajakumar, Alena Butryna, Matthew Lamm, Viktoriya Kuzmina, Joe Fenton, Aaron Cohen, Rachel Bernstein, Ray Kurzweil, Blaise Aguera-Arcas, Claire Cui, Marian Croak, Ed Chi and Quoc Le.
    • PaLM: Scaling Language Modeling with Pathways. [paper] [blog]

      Show AuthorsAakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez, Abhishek Rao, Parker Barnes, Yi Tay, Noam Shazeer, Vinodkumar Prabhakaran, Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradbury, Jacob Austin, Michael Isard, Guy Gur-Ari, Pengcheng Yin, Toju Duke, Anselm Levskaya, Sanjay Ghemawat, Sunipa Dev, Henryk Michalewski, Xavier Garcia, Vedant Misra, Kevin Robinson, Liam Fedus, Denny Zhou, Daphne Ippolito, David Luan, Hyeontaek Lim, Barret Zoph, Alexander Spiridonov, Ryan Sepassi, David Dohan, Shivani Agrawal, Mark Omernick, Andrew M. Dai, Thanumalayan Sankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz, Erica Moreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, Xuezhi Wang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei, Kathy Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov and Noah Fiedel.
    • UL2: Unifying Language Learning Paradigms. [paper] [blog]

      Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Xavier Garcia, Jason Wei, Xuezhi Wang, Hyung Won Chung, Dara Bahri, Tal Schuster, Huaixiu Steven Zheng, Denny Zhou, Neil Houlsby and Donald Metzler.

    • (Flan-T5 & Flan-PaLM & Flan-U-PaLM) Scaling Instruction-Finetuned Language Models. [paper] [blog]

      Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Xavier Garcia, Jason Wei, Xuezhi Wang, Hyung Won Chung, Dara Bahri, Tal Schuster, Huaixiu Steven Zheng, Denny Zhou, Neil Houlsby and Donald Metzler.

  • Meta

    • (Generative BST) Recipes for Building an Open-Domain Chatbot. [paper]

      Stephen Roller, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, Myle Ott, Kurt Shuster, Eric M. Smith, Y-Lan Boureau and Jason Weston.

    • (XGLM) Few-shot Learning with Multilingual Language Models. [paper]

      Xi Victoria Lin, Todor Mihaylov, Mikel Artetxe, Tianlu Wang, Shuohui Chen, Daniel Simig, Myle Ott, Naman Goyal, Shruti Bhosale, Jingfei Du, Ramakanth Pasunuru, Sam Shleifer, Punit Singh Koura, Vishrav Chaudhary, Brian O'Horo, Jeff Wang, Luke Zettlemoyer, Zornitsa Kozareva, Mona Diab, Veselin Stoyanov and Xian Li.

    • OPT: Open Pre-trained Transformer Language Models. [paper]

      Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, Todor Mihaylov, Myle Ott, Sam Shleifer, Kurt Shuster, Daniel Simig, Punit Singh Koura, Anjali Sridhar, Tianlu Wang and Luke Zettlemoyer.

    • BlenderBot 3: a deployed conversational agent that continually learns to responsibly engage. [paper] [blog]

      Kurt Shuster, Jing Xu, Mojtaba Komeili, Da Ju, Eric Michael Smith, Stephen Roller, Megan Ung, Moya Chen, Kushal Arora, Joshua Lane, Morteza Behrooz, William Ngan, Spencer Poff, Naman Goyal, Arthur Szlam, Y-Lan Boureau, Melanie Kambadur and Jason Weston.

    • Atlas: Few-shot Learning with Retrieval Augmented Language Models. [paper]

      Gautier Izacard, Patrick Lewis, Maria Lomeli, Lucas Hosseini, Fabio Petroni, Timo Schick, Jane Dwivedi-Yu, Armand Joulin, Sebastian Riedel and Edouard Grave.

    • LLaMA: Open and Efficient Foundation Language Models. [paper] [blog]

      Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave and Guillaume Lample.

  • BigScience

    • (T0++) Multitask Prompted Training Enables Zero-Shot Task Generalization. [paper]

      Show AuthorsVictor Sanh, Albert Webson, Colin Raffel, Stephen H. Bach, Lintang Sutawika, Zaid Alyafeai, Antoine Chaffin, Arnaud Stiegler, Teven Le Scao, Arun Raja, Manan Dey, M Saiful Bari, Canwen Xu, Urmish Thakker, Shanya Sharma Sharma, Eliza Szczechla, Taewoon Kim, Gunjan Chhablani, Nihal Nayak, Debajyoti Datta, Jonathan Chang, Mike Tian-Jian Jiang, Han Wang, Matteo Manica, Sheng Shen, Zheng Xin Yong, Harshit Pandey, Rachel Bawden, Thomas Wang, Trishala Neeraj, Jos Rozen, Abheesht Sharma, Andrea Santilli, Thibault Fevry, Jason Alan Fries, Ryan Teehan, Tali Bers, Stella Biderman, Leo Gao, Thomas Wolf and Alexander M. Rush.
    • BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. [paper]

      Show AuthorsBigScience Workshop: Teven Le Scao, Angela Fan, Christopher Akiki, Ellie Pavlick, Suzana Ilić, Daniel Hesslow, Roman Castagné, Alexandra Sasha Luccioni, François Yvon, Matthias Gallé, Jonathan Tow, Alexander M. Rush, Stella Biderman, Albert Webson, Pawan Sasanka Ammanamanchi, Thomas Wang, Benoît Sagot, Niklas Muennighoff, Albert Villanova del Moral, Olatunji Ruwase, Rachel Bawden, Stas Bekman, Angelina McMillan-Major, Iz Beltagy, Huu Nguyen, Lucile Saulnier, Samson Tan, Pedro Ortiz Suarez, Victor Sanh, Hugo Laurençon, Yacine Jernite, Julien Launay, Margaret Mitchell, Colin Raffel, Aaron Gokaslan, Adi Simhi, Aitor Soroa, Alham Fikri Aji, Amit Alfassy, Anna Rogers, Ariel Kreisberg Nitzav, Canwen Xu, Chenghao Mou, Chris Emezue, Christopher Klamm, Colin Leong, Daniel van Strien, David Ifeoluwa Adelani, Dragomir Radev, Eduardo González Ponferrada, Efrat Levkovizh, Ethan Kim, Eyal Bar Natan, Francesco De Toni, Gérard Dupont, Germán Kruszewski, Giada Pistilli, Hady Elsahar, Hamza Benyamina, Hieu Tran, Ian Yu, Idris Abdulmumin, Isaac Johnson, Itziar Gonzalez-Dios, Javier de la Rosa, Jenny Chim, Jesse Dodge, Jian Zhu, Jonathan Chang, Jörg Frohberg, Joseph Tobing, Joydeep Bhattacharjee, Khalid Almubarak, Kimbo Chen, Kyle Lo, Leandro Von Werra, Leon Weber, Long Phan, Loubna Ben allal, Ludovic Tanguy, Manan Dey, Manuel Romero Muñoz, Maraim Masoud, María Grandury, Mario Šaško, Max Huang, Maximin Coavoux, Mayank Singh, Mike Tian-Jian Jiang, Minh Chien Vu, Mohammad A. Jauhar, Mustafa Ghaleb, Nishant Subramani, Nora Kassner, Nurulaqilla Khamis, Olivier Nguyen, Omar Espejel, Ona de Gibert, Paulo Villegas , Peter Henderson, Pierre Colombo, Priscilla Amuok, Quentin Lhoest, Rheza Harliman, Rishi Bommasani, Roberto Luis López, Rui Ribeiro, Salomey Osei, Sampo Pyysalo, Sebastian Nagel, Shamik Bose, Shamsuddeen Hassan Muhammad, Shanya Sharma, Shayne Longpre, Somaieh Nikpoor, Stanislav Silberberg, Suhas Pai, Sydney Zink, Tiago Timponi Torrent, Timo Schick, Tristan Thrush, Valentin Danchev, Vassilina Nikoulina, Veronika Laippala, Violette Lepercq, Vrinda Prabhu, Zaid Alyafeai, Zeerak Talat, Arun Raja, Benjamin Heinzerling, Chenglei Si, Davut Emre Taşar, Elizabeth Salesky, Sabrina J. Mielke, Wilson Y. Lee, Abheesht Sharma, Andrea Santilli, Antoine Chaffin, Arnaud Stiegler, Debajyoti Datta, Eliza Szczechla, Gunjan Chhablani, Han Wang, Harshit Pandey, Hendrik Strobelt, Jason Alan Fries, Jos Rozen, Leo Gao, Lintang Sutawika, M Saiful Bari, Maged S. Al-shaibani, Matteo Manica, Nihal Nayak, Ryan Teehan, Samuel Albanie, Sheng Shen, Srulik Ben-David, Stephen H. Bach, Taewoon Kim, Tali Bers, Thibault Fevry, Trishala Neeraj, Urmish Thakker, Vikas Raunak, Xiangru Tang, Zheng-Xin Yong, Zhiqing Sun, Shaked Brody, Yallow Uri, Hadar Tojarieh, Adam Roberts, Hyung Won Chung, Jaesung Tae, Jason Phang, Ofir Press, Conglong Li, Deepak Narayanan, Hatim Bourfoune, Jared Casper, Jeff Rasley, Max Ryabinin, Mayank Mishra, Minjia Zhang, Mohammad Shoeybi, Myriam Peyrounette, Nicolas Patry, Nouamane Tazi, Omar Sanseviero, Patrick von Platen, Pierre Cornette, Pierre François Lavallée, Rémi Lacroix, Samyam Rajbhandari, Sanchit Gandhi, Shaden Smith, Stéphane Requena, Suraj Patil, Tim Dettmers, Ahmed Baruwa, Amanpreet Singh, Anastasia Cheveleva, Anne-Laure Ligozat, Arjun Subramonian, Aurélie Névéol, Charles Lovering, Dan Garrette, Deepak Tunuguntla, Ehud Reiter, Ekaterina Taktasheva, Ekaterina Voloshina, Eli Bogdanov, Genta Indra Winata, Hailey Schoelkopf, Jan-Christoph Kalo, Jekaterina Novikova, Jessica Zosa Forde, Jordan Clive, Jungo Kasai, Ken Kawamura, Liam Hazan, Marine Carpuat, Miruna Clinciu, Najoung Kim, Newton Cheng, Oleg Serikov, Omer Antverg, Oskar van der Wal, Rui Zhang, Ruochen Zhang, Sebastian Gehrmann, Shachar Mirkin, Shani Pais, Tatiana Shavrina, Thomas Scialom, Tian Yun, Tomasz Limisiewicz, Verena Rieser, Vitaly Protasov, Vladislav Mikhailov, Yada Pruksachatkun, Yonatan Belinkov, Zachary Bamberger, Zdeněk Kasner, Alice Rueda, Amanda Pestana, Amir Feizpour, Ammar Khan, Amy Faranak, Ana Santos, Anthony Hevia, Antigona Unldreaj, Arash Aghagol, Arezoo Abdollahi, Aycha Tammour, Azadeh HajiHosseini, Bahareh Behroozi, Benjamin Ajibade, Bharat Saxena, Carlos Muñoz Ferrandis, Danish Contractor, David Lansky, Davis David, Douwe Kiela, Duong A. Nguyen, Edward Tan, Emi Baylor, Ezinwanne Ozoani, Fatima Mirza, Frankline Ononiwu, Habib Rezanejad, Hessie Jones, Indrani Bhattacharya, Irene Solaiman, Irina Sedenko, Isar Nejadgholi, Jesse Passmore, Josh Seltzer, Julio Bonis Sanz, Livia Dutra, Mairon Samagaio, Maraim Elbadri, Margot Mieskes, Marissa Gerchick, Martha Akinlolu, Michael McKenna, Mike Qiu, Muhammed Ghauri, Mykola Burynok, Nafis Abrar, Nazneen Rajani, Nour Elkott, Nour Fahmy, Olanrewaju Samuel, Ran An, Rasmus Kromann, Ryan Hao, Samira Alizadeh, Sarmad Shubber, Silas Wang, Sourav Roy, Sylvain Viguier, Thanh Le, Tobi Oyebade, Trieu Le, Yoyo Yang, Zach Nguyen, Abhinav Ramesh Kashyap, Alfredo Palasciano, Alison Callahan, Anima Shukla, Antonio Miranda-Escalada, Ayush Singh, Benjamin Beilharz, Bo Wang, Caio Brito, Chenxi Zhou, Chirag Jain, Chuxin Xu, Clémentine Fourrier, Daniel León Periñán, Daniel Molano, Dian Yu, Enrique Manjavacas, Fabio Barth, Florian Fuhrimann, Gabriel Altay, Giyaseddin Bayrak, Gully Burns, Helena U. Vrabec, Imane Bello, Ishani Dash, Jihyun Kang, John Giorgi, Jonas Golde, Jose David Posada, Karthik Rangasai Sivaraman, Lokesh Bulchandani, Lu Liu, Luisa Shinzato, Madeleine Hahn de Bykhovetz, Maiko Takeuchi, Marc Pàmies, Maria A Castillo, Marianna Nezhurina, Mario Sänger, Matthias Samwald, Michael Cullan, Michael Weinberg, Michiel De Wolf, Mina Mihaljcic, Minna Liu, Moritz Freidank, Myungsun Kang, Natasha Seelam, Nathan Dahlberg, Nicholas Michio Broad, Nikolaus Muellner, Pascale Fung, Patrick Haller, Ramya Chandrasekhar, Renata Eisenberg, Robert Martin, Rodrigo Canalli, Rosaline Su, Ruisi Su, Samuel Cahyawijaya, Samuele Garda, Shlok S Deshmukh, Shubhanshu Mishra, Sid Kiblawi, Simon Ott, Sinee Sang-aroonsiri, Srishti Kumar, Stefan Schweter, Sushil Bharati, Tanmay Laud, Théo Gigant, Tomoya Kainuma, Wojciech Kusa, Yanis Labrak, Yash Shailesh Bajaj, Yash Venkatraman, Yifan Xu, Yingxin Xu, Yu Xu, Zhe Tan, Zhongli Xie, Zifan Ye, Mathilde Bras, Younes Belkada and Thomas Wolf.
  • Amazon

    • AlexaTM 20B: Few-Shot Learning Using a Large-Scale Multilingual Seq2Seq Model. [paper] [blog]

      Saleh Soltan, Shankar Ananthakrishnan, Jack FitzGerald, Rahul Gupta, Wael Hamza, Haidar Khan, Charith Peris, Stephen Rawls, Andy Rosenbaum, Anna Rumshisky, Chandana Satya Prakash, Mukund Sridhar, Fabian Triefenbach, Apurv Verma, Gokhan Tur and Prem Natarajan.

  • AI21 Labs

    • Jurassic-1: Technical Details and Evaluation. [paper] [blog]

      Opher Lieber, Or Sharir, Barak Lenz and Yoav Shoham.

  • Baidu

    • ERNIE 3.0: Large-scale Knowledge Enhanced Pre-training for Language Understanding and Generation. [paper]

      Yu Sun, Shuohuan Wang, Shikun Feng, Siyu Ding, Chao Pang, Junyuan Shang, Jiaxiang Liu, Xuyi Chen, Yanbin Zhao, Yuxiang Lu, Weixin Liu, Zhihua Wu, Weibao Gong, Jianzhong Liang, Zhizhou Shang, Peng Sun, Wei Liu, Xuan Ouyang, Dianhai Yu, Hao Tian, Hua Wu and Haifeng Wang.

    • ERNIE 3.0 Titan: Exploring Larger-scale Knowledge Enhanced Pre-training for Language Understanding and Generation. [paper]

      Show AuthorsShuohuan Wang, Yu Sun, Yang Xiang, Zhihua Wu, Siyu Ding, Weibao Gong, Shikun Feng, Junyuan Shang, Yanbin Zhao, Chao Pang, Jiaxiang Liu, Xuyi Chen, Yuxiang Lu, Weixin Liu, Xi Wang, Yangfan Bai, Qiuliang Chen, Li Zhao, Shiyong Li, Peng Sun, Dianhai Yu, Yanjun Ma, Hao Tian, Hua Wu, Tian Wu, Wei Zeng, Ge Li, Wen Gao and Haifeng Wang.
    • PLATO-XL: Exploring the Large-scale Pre-training of Dialogue Generation. [paper]

      Siqi Bao, Huang He, Fan Wang, Hua Wu, Haifeng Wang, Wenquan Wu, Zhihua Wu, Zhen Guo, Hua Lu, Xinxian Huang, Xin Tian, Xinchao Xu, Yingzhan Lin and Zheng-Yu Niu.

    • ERNIE Bot [news]

  • Anthropic

    • (Anthropic-52) Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback [paper]

      Show AuthorsYuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, Nicholas Joseph, Saurav Kadavath, Jackson Kernion, Tom Conerly, Sheer El-Showk, Nelson Elhage, Zac Hatfield-Dodds, Danny Hernandez, Tristan Hume, Scott Johnston, Shauna Kravec, Liane Lovitt, Neel Nanda, Catherine Olsson, Dario Amodei, Tom Brown, Jack Clark, Sam McCandlish, Chris Olah, Ben Mann and Jared Kaplan.
    • (Claude) Constitutional AI: Harmlessness from AI Feedback. [paper]

      Show AuthorsYuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Olsson, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, Ethan Perez, Jamie Kerr, Jared Mueller, Jeffrey Ladish, Joshua Landau, Kamal Ndousse, Kamile Lukosuite, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noemi Mercado, Nova DasSarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna Kravec, Sheer El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Conerly, Tom Henighan, Tristan Hume, Samuel R. Bowman, Zac Hatfield-Dodds, Ben Mann, Dario Amodei, Nicholas Joseph, Sam McCandlish, Tom Brown and Jared Kaplan.
  • EleutherAI

    • GPT-J-6B: 6B JAX-Based Transformer. [blog]

    • GPT-NeoX-20B: An Open-Source Autoregressive Language Model. [paper] [blog]

      Sid Black, Stella Biderman, Eric Hallahan, Quentin Anthony, Leo Gao, Laurence Golding, Horace He, Connor Leahy, Kyle McDonell, Jason Phang, Michael Pieler, USVSN Sai Prashanth, Shivanshu Purohit, Laria Reynolds, Jonathan Tow, Ben Wang, Samuel Weinbach.

  • Tsinghua University

    • CPM: A Large-scale Generative Chinese Pre-trained Language Model. [paper]

      Zhengyan Zhang, Xu Han, Hao Zhou, Pei Ke, Yuxian Gu, Deming Ye, Yujia Qin, Yusheng Su, Haozhe Ji, Jian Guan, Fanchao Qi, Xiaozhi Wang, Yanan Zheng, Guoyang Zeng, Huanqi Cao, Shengqi Chen, Daixuan Li, Zhenbo Sun, Zhiyuan Liu, Minlie Huang, Wentao Han, Jie Tang, Juanzi Li, Xiaoyan Zhu, Maosong Sun.

    • CPM-2: Large-scale Cost-effective Pre-trained Language Models. [paper]

      Zhengyan Zhang, Yuxian Gu, Xu Han, Shengqi Chen, Chaojun Xiao, Zhenbo Sun, Yuan Yao, Fanchao Qi, Jian Guan, Pei Ke, Yanzheng Cai, Guoyang Zeng, Zhixing Tan, Zhiyuan Liu, Minlie Huang, Wentao Han, Yang Liu, Xiaoyan Zhu, Maosong Sun.

    • Chinese-Transformer-XL [repo]

    • EVA: An Open-Domain Chinese Dialogue System with Large-Scale Generative Pre-Training. [paper]

      Hao Zhou, Pei Ke, Zheng Zhang, Yuxian Gu, Yinhe Zheng, Chujie Zheng, Yida Wang, Chen Henry Wu, Hao Sun, Xiaocong Yang, Bosi Wen, Xiaoyan Zhu, Minlie Huang, Jie Tang.

    • EVA2.0: Investigating Open-Domain Chinese Dialogue Systems with Large-Scale Pre-Training. [paper]

      Yuxian Gu, Jiaxin Wen, Hao Sun, Yi Song, Pei Ke, Chujie Zheng, Zheng Zhang, Jianzhu Yao, Xiaoyan Zhu, Jie Tang, Minlie Huang.

    • GLM: General Language Model Pretraining with Autoregressive Blank Infilling. [paper]

      Zhengxiao Du, Yujie Qian, Xiao Liu, Ming Ding, Jiezhong Qiu, Zhilin Yang, Jie Tang.

    • GLM-130B: An Open Bilingual Pre-trained Model. [paper]

      Aohan Zeng, Xiao Liu, Zhengxiao Du, Zihan Wang, Hanyu Lai, Ming Ding, Zhuoyi Yang, Yifan Xu, Wendi Zheng, Xiao Xia, Weng Lam Tam, Zixuan Ma, Yufei Xue, Jidong Zhai, Wenguang Chen, Zhiyuan Liu, Peng Zhang, Yuxiao Dong and Jie Tang.

    • ChatGLM. [blog]

  • IDEA

    • Fengshenbang 1.0: Being the Foundation of Chinese Cognitive Intelligence. [paper] [blog]

      Junjie Wang, Yuxiang Zhang, Lin Zhang, Ping Yang, Xinyu Gao, Ziwei Wu, Xiaoqun Dong, Junqing He, Jianheng Zhuo, Qi Yang, Yongfeng Huang, Xiayu Li, Yanghan Wu, Junyu Lu, Xinyu Zhu, Weifeng Chen, Ting Han, Kunhao Pan, Rui Wang, Hao Wang, Xiaojun Wu, Zhongshen Zeng, Chongpei Chen, Ruyi Gan and Jiaxing Zhang.

      • Wenzhong
      • Randeng
      • Yuyuan
      • Zhouwenwang
  • Huawei

    • PanGu-α: Large-scale Autoregressive Pretrained Chinese Language Models with Auto-parallel Computation. [paper]

      Show AuthorsWei Zeng, Xiaozhe Ren, Teng Su, Hui Wang, Yi Liao, Zhiwei Wang, Xin Jiang, ZhenZhang Yang, Kaisheng Wang, Xiaoda Zhang, Chen Li, Ziyan Gong, Yifan Yao, Xinjing Huang, Jun Wang, Jianfeng Yu, Qi Guo, Yue Yu, Yan Zhang, Jin Wang, Hengtao Tao, Dasen Yan, Zexuan Yi, Fang Peng, Fangqing Jiang, Han Zhang, Lingfeng Deng, Yehong Zhang, Zhe Lin, Chao Zhang, Shaojie Zhang, Mingyue Guo, Shanzhi Gu, Gaojun Fan, Yaowei Wang, Xuefeng Jin, Qun Liu and Yonghong Tian.
    • PanGu-Bot: Efficient Generative Dialogue Pre-training from Pre-trained Language Model. [paper]

      Fei Mi, Yitong Li, Yulong Zeng, Jingyan Zhou, Yasheng Wang, Chuanfei Xu, Lifeng Shang, Xin Jiang, Shiqi Zhao and Qun Liu.

  • Inspur

    • Yuan 1.0: Large-Scale Pre-trained Language Model in Zero-Shot and Few-Shot Learning. [paper] [blog]

      Shaohua Wu, Xudong Zhao, Tong Yu, Rongguo Zhang, Chong Shen, Hongli Liu, Feng Li, Hong Zhu, Jiangang Luo, Liang Xu and Xuanwei Zhang.

  • WeChat AI

    • WeLM: A Well-Read Pre-trained Language Model for Chinese. [paper]

      Hui Su, Xiao Zhou, Houjin Yu, Yuwen Chen, Zilin Zhu, Yang Yu and Jie Zhou.

Analysis

Preview

Comprehensive

  • Holistic Evaluation of Language Models. [paper] [blog]

    Show AuthorsPercy Liang, Rishi Bommasani, Tony Lee, Dimitris Tsipras, Dilara Soylu, Michihiro Yasunaga, Yian Zhang, Deepak Narayanan, Yuhuai Wu, Ananya Kumar, Benjamin Newman, Binhang Yuan, Bobby Yan, Ce Zhang, Christian Cosgrove, Christopher D. Manning, Christopher Ré, Diana Acosta-Navas, Drew A. Hudson, Eric Zelikman, Esin Durmus, Faisal Ladhak, Frieda Rong, Hongyu Ren, Huaxiu Yao, Jue Wang, Keshav Santhanam, Laurel Orr, Lucia Zheng, Mert Yuksekgonul, Mirac Suzgun, Nathan Kim, Neel Guha, Niladri Chatterji, Omar Khattab, Peter Henderson, Qian Huang, Ryan Chi, Sang Michael Xie, Shibani Santurkar, Surya Ganguli, Tatsunori Hashimoto, Thomas Icard, Tianyi Zhang, Vishrav Chaudhary, William Wang, Xuechen Li, Yifan Mai, Yuhui Zhang and Yuta Koreeda.

    Benchmark 30 prominent language models across a wide range of scenarios and for a broad range of metrics to elucidate their capabilities and risks.

  • On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜 [paper]

    Emily M. Bender, Timnit Gebru, Angelina McMillan-Major and Shmargaret Shmitchell.

    Identified a wide variety of costs and risks associated with the rush for ever larger LMs, including: environmental costs, financial costs, opportunity cost and the risk of substantial harms.

  • TruthfulQA: Measuring How Models Mimic Human Falsehoods. [paper]

    Stephanie Lin, Jacob Hilton and Owain Evans.

    Models generated many false answers that mimic popular misconceptions and have the potential to deceive humans. The largest models were generally the least truthful.

  • How Much Knowledge Can You Pack Into the Parameters of a Language Model? [paper]

    Adam Roberts, Colin Raffel and Noam Shazeer.

    The maximum-likelihood objective used to train our model provides no guarantees as to whether a model will learn a fact or not. This makes it difficult to ensure that the model obtains specific knowledge over the course of pre-training and prevents us from explicitly updating or removing knowledge from a pre-trained model.

  • On the Opportunities and Risks of Foundation Models. [paper]

    Show AuthorsRishi Bommasani, Drew A. Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri Chatterji, Annie Chen, Kathleen Creel, Jared Quincy Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang Wei Koh, Mark Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher D. Manning, Suvir Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Ben Newman, Allen Nie, Juan Carlos Niebles, Hamed Nilforoshan, Julian Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon Sung Park, Chris Piech, Eva Portelance, Christopher Potts, Aditi Raghunathan, Rob Reich, Hongyu Ren, Frieda Rong, Yusuf Roohani, Camilo Ruiz, Jack Ryan, Christopher Ré, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishnan Srinivasan, Alex Tamkin, Rohan Taori, Armin W. Thomas, Florian Tramèr, Rose E. Wang, William Wang , Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou and Percy Liang.

    This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities and technical principles to their applications and societal impact.

Hallucination & Disinformation

  • Survey of Hallucination in Natural Language Generation. [paper]

    Show AuthorsPercy Liang, Rishi Bommasani, Tony Lee, Dimitris Tsipras, Dilara Soylu, Michihiro Yasunaga, Yian Zhang, Deepak Narayanan, Yuhuai Wu, Ananya Kumar, Benjamin Newman, Binhang Yuan, Bobby Yan, Ce Zhang, Christian Cosgrove, Christopher D. Manning, Christopher Ré, Diana Acosta-Navas, Drew A. Hudson, Eric Zelikman, Esin Durmus, Faisal Ladhak, Frieda Rong, Hongyu Ren, Huaxiu Yao, Jue Wang, Keshav Santhanam, Laurel Orr, Lucia Zheng, Mert Yuksekgonul, Mirac Suzgun, Nathan Kim, Neel Guha, Niladri Chatterji, Omar Khattab, Peter Henderson, Qian Huang, Ryan Chi, Sang Michael Xie, Shibani Santurkar, Surya Ganguli, Tatsunori Hashimoto, Thomas Icard, Tianyi Zhang, Vishrav Chaudhary, William Wang, Xuechen Li, Yifan Mai, Yuhui Zhang and Yuta Koreeda.

    Benchmark 30 prominent language models across a wide range of scenarios and for a broad range of metrics to elucidate their capabilities and risks.

  • On the Origin of Hallucinations in Conversational Models: Is it the Datasets or the Models? [paper]

    Nouha Dziri, Sivan Milton, Mo Yu, Osmar Zaiane and Siva Reddy.

    Reveals that the standard benchmarks consist of >60% hallucinated responses, leading to models that not only hallucinate but even amplify hallucinations..

  • Generative Language Models and Automated Influence Operations: Emerging Threats and Potential Mitigations. [paper]

    Josh A. Goldstein, Girish Sastry, Micah Musser, Renee DiResta, Matthew Gentzel and Katerina Sedova.

    This report aims to assess: how might language models change influence operations, and what steps can be taken to mitigate these threats? This task is inherently speculative, as both AI and influence operations are changing quickly.

  • Forecasting Potential Misuses of Language Models for Disinformation Campaigns—and How to Reduce Risk. [blog]

    OpenAI researchers collaborated with Georgetown University’s Center for Security and Emerging Technology and the Stanford Internet Observatory to investigate how large language models might be misused for disinformation purposes.

  • ChatGPT Wrote a Terrible Gizmodo Article. [blog]

    ChatGPT kept including incorrect information in its explainer—sometimes mixing up basic facts about the history of its own technology.

  • Temporary policy: ChatGPT is banned. [blog]

    While the answers which ChatGPT produces have a high rate of being incorrect, they typically look like they might be good and the answers are very easy to produce.

Bias & Toxicity

  • On Second Thought, Let's Not Think Step by Step! Bias and Toxicity in Zero-Shot Reasoning. [paper]

    Omar Shaikh, Hongxin Zhang, William Held, Michael Bernstein and Diyi Yang.

    Find that using zero-shot CoT reasoning in a prompt can significantly increase a model’s likelihood to produce undesirable output.

  • The Woman Worked as a Babysitter: On Biases in Language Generation. [paper]

    Emily Sheng, Kai-Wei Chang, Premkumar Natarajan and Nanyun Peng.

    A systematic study of biases in natural language generation (NLG) by analyzing text generated from prompts that contain mentions of different demographic groups.

  • RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. [paper]

    Samuel Gehman, Suchin Gururangan, Maarten Sap, Yejin Choi and Noah A. Smith.

    Using RealToxicityPrompts, we find that pretrained LMs can degenerate into toxic text even from seemingly innocuous prompts.

  • OpenAI's new ChatGPT bot: 10 dangerous things it's capable of. [blog]

    As the erudite machinery turns into a viral sensation, humans have started to discover some of the AI's biases, like the desire to wipe out humanity.

Security Risk

  • OpwnAI: AI That Can Save the Day or HACK it Away. [blog]

    AI models can be used to create a full infection flow, from spear-phishing to running a reverse shell.

  • OpwnAI : Cybercriminals Starting to Use ChatGPT. [blog]

    There are already instances of cybercriminals using OpenAI to develop malicious tools.

  • Security risks of ChatGPT and other AI text generators. [blog]

    Successfully asked ChatGPT to write a convincing phishing email and to create JavaScript that could be used to steal personal information.

  • The security threat of AI-powered cyberattacks. [paper]

    Matti Aksela, Samuel Marchal, Andrew Patel, Lina Rosenstedt and WithSecure.

    Investigate the security threat of AI-enabled cyberattacks by summarising current knowledge on the topic.

  • How hackers might be exploiting ChatGPT. [blog]

    The viral AI chatbot ChatGPT might advise threat actors how to hack into networks with ease.

LM Attack

  • Extracting Training Data from Large Language Models. [paper]

    Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, Alina Oprea and Colin Raffel.

    This paper demonstrates that an adversary can perform a training data extraction attack to recover individual training examples by querying the language model.

  • Ignore Previous Prompt: Attack Techniques For Language Models. [paper]

    Fábio Perez and Ian Ribeiro.

    Study prompt injection attacks against LLMs and propose a framework to explore such attacks; Investigate two specific attacks: goal hijacking and prompt leaking.

Environment

  • Carbon Emissions and Large Neural Network Training. [paper]

    David Patterson, Joseph Gonzalez, Quoc Le, Chen Liang, Lluis-Miquel Munguia, Daniel Rothchild, David So, Maud Texier and Jeff Dean.

    Calculate the energy use and carbon footprint of several recent large models-T5, Meena, GShard, Switch Transformer, and GPT-3.

Detection

Papers

Survey

  • Automatic Detection of Machine Generated Text: A Critical Survey. [paper]

    Ganesh Jawahar, Muhammad Abdul-Mageed and Laks V.S. Lakshmanan.

  • 在线社交网络文本内容对抗技术. [paper]

    刘晓明, 张兆晗, 杨晨阳, 张宇辰, 沈超, 周亚东 and 管晓宏.

  • Synthetic Text Detection: Systemic Literature Review. [paper]

    Jesus Guerrero, Izzat Alsmadi.

  • Machine Generated Text: A Comprehensive Survey of Threat Models and Detection Methods. [paper]

    Evan Crothers, Nathalie Japkowicz and Herna Viktor.

  • Deepfake Text Detection: Limitations and Opportunities. [paper]

    Jiameng Pu, Zain Sarwar, Sifat Muhammad Abdullah, Abdullah Rehman, Yoonjin Kim, Parantapa Bhattacharya, Mobin Javed and Bimal Viswanath.

  • The Science of Detecting LLM-Generated Texts. [paper]

    Ruixiang Tang, Yu-Neng Chuang and Xia Hu.

  • To ChatGPT, or not to ChatGPT: That is the question! [paper]

    Alessandro Pegoraro, Kavita Kumari, Hossein Fereidooni and Ahmad-Reza Sadeghi.

  • The Age of Synthetic Realities: Challenges and Opportunities. [paper]

    João Phillipe Cardenuto, Jing Yang, Rafael Padilha, Renjie Wan, Daniel Moreira, Haoliang Li, Shiqi Wang, Fernanda Andaló, Sébastien Marcel and Anderson Rocha.

  • Attribution and Obfuscation of Neural Text Authorship: A Data Mining Perspective. [paper]

    Adaku Uchendu, Thai Le and Dongwon Lee.

  • Detecting Artificial Intelligence: A New Cyberarms Race Begins. [paper]

    Mark Campbell and Mlađan Jovanović.

  • Detecting ChatGPT: A Survey of the State of Detecting ChatGPT-Generated Text. [paper]

    Mahdi Dhaini, Wessel Poelman and Ege Erdogan.

  • A Survey on LLM-gernerated Text Detection: Necessity, Methods, and Future Directions. [paper]

    Junchao Wu, Shu Yang, Runzhe Zhan, Yulin Yuan, Derek F. Wong and Lidia S. Chao.

  • Towards Possibilities & Impossibilities of AI-generated Text Detection: A Survey. [paper]

    Soumya Suvra Ghosal, Souradip Chakraborty, Jonas Geiping, Furong Huang, Dinesh Manocha and Amrit Singh Bedi.

  • A Survey on Detection of LLMs-Generated Content. [paper]

    Xianjun Yang, Liangming Pan, Xuandong Zhao, Haifeng Chen, Linda Petzold, William Yang Wang and Wei Cheng.

  • A Survey of Text Watermarking in the Era of Large Language Models. [paper]

    Aiwei Liu, Leyi Pan, Yijian Lu, Jingjing Li, Xuming Hu, Lijie Wen, Irwin King and Philip S. Yu.

  • Detection of Machine-Generated Text: Literature Survey. [paper]

    Dmytro Valiaiev.

  • Detecting Multimedia Generated by Large AI Models: A Survey. [paper]

    Li Lin, Neeraj Gupta, Yue Zhang, Hainan Ren, Chun-Hao Liu, Feng Ding, Xin Wang, Xin Li, Luisa Verdoliva and Shu Hu.

  • Copyright Protection in Generative AI: A Technical Perspective. [paper]

    Jie Ren, Han Xu, Pengfei He, Yingqian Cui, Shenglai Zeng, Jiankun Zhang, Hongzhi Wen, Jiayuan Ding, Hui Liu, Yi Chang and Jiliang Tang.

  • Beyond the Human Eye: Comprehensive Approaches to AI Text Detection. [paper]

    Jamal Goddard, Yuksel Celik and Sanjay Goel.

  • A Survey of AI-generated Text Forensic Systems: Detection, Attribution, and Characterization. [paper]

    Tharindu Kumarage, Garima Agrawal, Paras Sheth, Raha Moraffah, Aman Chadha, Joshua Garland and Huan Liu.

  • Decoding the AI Pen: Techniques and Challenges in Detecting AI-Generated Text. [paper]

    Sara Abdali, Richard Anarfi, CJ Barberan and Jia He.

Human Detection

  • Deepfake Bot Submissions to Federal Public Comment Websites Cannot Be Distinguished from Human Submissions. [paper]

    Max Weiss.

    Generate and submit 1,001 deepfake comments regarding a Medicaid reform waiver to a federal public comment website. The human classification results are no better than would have been gotten by random guessing.

  • RoFT: A Tool for Evaluating Human Detection of Machine-Generated Text. [paper]

    Liam Dugan, Daphne Ippolito, Arun Kirubarajan and Chris Callison-Burch.

    Develop the Real or Fake Text (RoFT) system, a novel application for simultaneously collecting quality annotations of machinegenerated text while allowing the public to assess and improve their skill at detecting machinegenerated text.

  • All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text. [paper]

    Elizabeth Clark, Tal August, Sofia Serrano, Nikita Haduong, Suchin Gururangan and Noah A. Smith.

    Assess nonexperts’ ability to distinguish between humanand machine-authored text, and test three evaluator-training methods to see if we could improve people’s ability to identify machinegenerated text.

  • Human Heuristics for AI-Generated Language Are Flawed. [paper]

    Maurice Jakesch, Jeffrey Hancock and Mor Naaman.

    Study how humans discern whether verbal self-presentations were generated by AI. Show that human judgments of AI-generated language are handicapped by intuitive but flawed heuristics.

  • Real or Fake Text?: Investigating Human Ability to Detect Boundaries Between Human-Written and Machine-Generated Text. [paper]

    Liam Dugan, Daphne Ippolito, Arun Kirubarajan, Sherry Shi and Chris Callison-Burch.

    Analyze how a variety of variables (model size, decoding strategy, fine-tuning, prompt genre, etc.) affect human detection performance.

  • AI model GPT-3 (dis)informs us better than humans. [paper]

    Giovanni Spitale, Nikola Biller-Andorno and Federico Germani.

    Evaluate whether recruited individuals can distinguish disinformation from accurate information, and determine whether a tweet has been written by a Twitter user or by GPT-3.

  • Creating a Large Language Model of a Philosopher. [paper]

    Eric Schwitzgebel, David Schwitzgebel and Anna Strasser.

    Fine-tuned the GPT-3 on the corpus of Daniel Dennett, then asked it a series of philosophical questions. Ordinary research participants untrained in philosophy were at or near chance in distinguishing GPT-3’s answers from those of an “actual human philosopher”. .

  • Does Human Collaboration Enhance the Accuracy of Identifying LLM-Generated Deepfake Texts? [paper]

    Adaku Uchendu, Jooyoung Lee, Hua Shen, Thai Le, Ting-Hao 'Kenneth' Huang and Dongwon Lee.

    Find that: (1) expert humans detect deepfake texts significantly better than non-expert humans, (2) synchronous teams on the Upwork detect deepfake texts significantly better than individuals, while asynchronous teams on the AMT detect deepfake texts weakly better than individuals, and (3) among various error categories, examining coherence and consistency in texts is useful in detecting deepfake texts.

  • Too Good to Be True? An Empirical Study of ChatGPT Capabilities for Academic Writing and Implications for Academic Misconduct. [paper]

    Peter André Busch and Geir Inge Hausvik.

    Ask 15 faculty members to assess a total of ten answers (two each) to an actual IS exam question, of which students wrote five, and ChatGPT generated five. Find that ChatGPT can generate answers of generally good quality that may pass as human-written text by examiners.

  • Game of Tones: Faculty detection of GPT-4 generated content in university assessments. [paper]

    Mike Perkins, Jasper Roe, Darius Postma, James McGaughran and Don Hickerson.

    This study explores the robustness of university assessments against the use of Open AI's Generative Pre-Trained Transformer 4 (GPT-4) generated content and evaluates the ability of academic staff to detect its use when supported by the Turnitin Artificial Intelligence (AI) detection tool.

  • The Unseen A+ Student: Navigating the Impact of Large Language Models in the Classroom. [paper]

    Matyas Bohacek.

    Collect a dataset of authentic high school coursework and generate their AI alternatives and text continuations using ChatGPT with 4.0, 3.5, and 3.5 Legacy backbones. Through a study involving student peers, we found that ChatGPT can quickly produce high-school-level coursework that peers consider better than human-written text, even in a low-resourced language like Czech. Moreover, show that the AI text detectors fail to identify these texts in Czech.

  • Can linguists distinguish between ChatGPT/AI and human writing?: A study of research ethics and academic publishing. [paper]

    J. Elliott Casal and Matt Kessler.

    Investigate: 1) the extent to which linguists/reviewers from top journals can distinguish AI- from human-generated writing, 2) what the basis of reviewers’ decisions are, and 3) the extent to which editors of top Applied Linguistics journals believe AI tools are ethical for research purposes.

  • Detection of GPT-4 Generated Text in Higher Education: Combining Academic Judgement and Software to Identify Generative AI Tool Misuse. [paper]

    Mike Perkins, Jasper Roe, Darius Postma, James McGaughran and Don Hickerson.

    Explore the capability of academic staff assisted by the Turnitin Artificial Intelligence (AI) detection tool to identify the use of AI-generated content in university assessments.

  • ChatGPT versus Human Essayists: An Exploration of the Impact of Artificial Intelligence for Authorship and Academic Integrity in the Humanities. [paper]

    Tom Revell, Will Yeadon, Glenn Cahilly-Bretzin, Isabella Clarke, George Manning, Jasmine Jones, Clare Mulley, Rafael Pascual, Natasha Bradley, Daniel Thomas and Francis Leneghan.

    Examine AI’s ability to write essays analysing Old English poetry; human markers assessed and attempted to distinguish them from authentic analyses of poetry by first-year undergraduate students in English at the University of Oxford.

  • Evaluating AI and Human Authorship Quality in Academic Writing through Physics Essays. [paper]

    Will Yeadon, Elise Agra, Oto-obong Inyang, Paul Mackay and Arin Mizouri.

    Evaluate n=300 short-form physics essay submissions, equally divided between student work submitted before the introduction of ChatGPT and those generated by OpenAI's GPT-4. In blinded evaluations conducted by five independent markers who were unaware of the origin of the essays, we observed no statistically significant differences in scores between essays authored by humans and those produced by AI (p-value =0.107, α = 0.05).

  • A feasibility study for the application of AI-generated conversations in pragmatic analysis. [paper]

    Xi Chen, Jun Li and Yuting Ye.

    Compare 148 ChatGPT-generated conversations with 82 human-written ones and 354 human evaluations of these conversations. The data are analysed using various methods, including traditional speech strategy coding, four computational methods developed in NLP, and four statistical tests.

Automatic Detection

  • Defending Against Neural Fake News. [paper]

    Rowan Zellers, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner and Yejin Choi.

    Present a model for controllable text generation called Grover. Find that the best way to detect neural fake news is to use a model that is also a generator.

  • GLTR: Statistical Detection and Visualization of Generated Text. [paper]

    Sebastian Gehrmann, Hendrik Strobelt and Alexander M. Rush.

    A Giant Language model Test Room. GLTR aims to both teach users what to be aware of when assessing whether a text is real, and to assist them in performing forensic analyses.

  • Real or Fake? Learning to Discriminate Machine from Human Generated Text. [paper]

    Anton Bakhtin, Sam Gross, Myle Ott, Yuntian Deng, Marc'Aurelio Ranzato and Arthur Szlam.

    Use Energy-based models to discriminate real text from text generated by the auto-regressive models.

  • The Limitations of Stylometry for Detecting Machine-Generated Fake News. [paper]

    Tal Schuster, Roei Schuster, Darsh J Shah and Regina Barzilay.

    Examine the state-of-the-art stylometry model, and find it effective in preventing impersonation, but limited in detecting LM-generated misinformation.

  • Automatic Detection of Generated Text is Easiest when Humans are Fooled. [paper]

    Daphne Ippolito, Daniel Duckworth, Chris Callison-Burch and Douglas Eck.

    A comprehensive study of generated text detection systems’ sensitivity to model structure, decoding strategy, and excerpt length.

  • Reverse Engineering Configurations of Neural Text Generation Models. [paper]

    Yi Tay, Dara Bahri, Che Zheng, Clifford Brunk, Donald Metzler and Andrew Tomkins.

    Propose the new task of distinguishing which of several variants (e.g., sampling methods, top-k probabilities, model architectures, etc.) of a given model generated some piece of text. Find that detectable artifacts are present and that different modeling choices can be inferred by looking at generated text alone.

  • TweepFake: about Detecting Deepfake Tweets. [paper]

    Tiziano Fagni, Fabrizio Falchi, Margherita Gambini, Antonio Martella and Maurizio Tesconi.

    Collect the first dataset of real deepfake tweets, TweepFake. Evaluate 13 deepfake text detection methods.

  • Identifying Automatically Generated Headlines using Transformers. [paper]

    Antonis Maronikolakis, Hinrich Schutze and Mark Stevenson.

    Create a dataset containing human and computer-generated headlines. Transformers achieved an overall accuracy of 85.7%.

  • Detecting Cross-Modal Inconsistency to Defend Against Neural Fake News. [paper]

    Reuben Tan, Bryan A. Plummer and Kate Saenko.

    Present DIDAN, a approach which exploits possible semantic inconsistencies between the text and image/captions to detect machine-generated articles.

  • Neural Deepfake Detection with Factual Structure of Text. [paper]

    Wanjun Zhong, Duyu Tang, Zenan Xu, Ruize Wang, Nan Duan, Ming Zhou, Jiahai Wang and Jian Yin.

    Propose a graph-based model that utilizes the factual structure of a document for deepfake detection of text.

  • Authorship Attribution for Neural Text Generation. [paper]

    Adaku Uchendu, Thai Le, Kai Shu and Dongwon Lee.

    Investigate the authorship attribution problem in three versions: (1) given two texts T1 and T2, are both generated by the same method or not? (2) is the given text T written by a human or machine? (3) given a text T and k candidate neural methods, can we single out the method (among k alternatives) that generated T?

  • How Effectively Can Machines Defend Against Machine-Generated Fake News? An Empirical Study. [paper]

    Meghana Moorthy Bhat and Srinivasan Parthasarathy.

    Measure the performance of models by looking at accuracy with respect to perturbations introduced in this work. Find that success of style-based classifiers are limited when real articles are perturbed even under extreme modifications.

  • Feature-based detection of automated language models: tackling GPT-2, GPT-3 and Grover. [paper]

    Leon Fröhling and Arkaitz Zubiaga.

    A feature-based detection approach relies on features that discriminate between human and machine text by modelling properties and dimensions in which both types of text differ.

  • Detecting Bot-Generated Text by Characterizing Linguistic Accommodation in Human-Bot Interactions. [paper]

    Paras Bhatt and Anthony Rios.

    Show that bot-generated text detection methods are more robust across datasets and models if we use information about how people respond to it rather than using the bot’s text directly.

  • Artificial Text Detection via Examining the Topology of Attention Maps. [paper]

    Laida Kushnareva, Daniil Cherniavskii, Vladislav Mikhailov, Ekaterina Artemova, Serguei Barannikov, Alexander Bernstein, Irina Piontkovskaya, Dmitri Piontkovski and Evgeny Burnaev.

    Propose three novel types of interpretable topological features based on Topological Data Analysis (TDA). The features derived from the BERT model outperform count- and neural-based baselines up to 10% on three common datasets, and tend to be the most robust towards unseen GPT-style generation models as opposed to existing methods.

  • Unsupervised and Distributional Detection of Machine-Generated Text. [paper]

    Matthias Gallé, Jos Rozen, Germán Kruszewski and Hady Elsahar.

    Propose a method to detect those machine-generated documents leveraging repeated higher-order n-grams, which we show over-appear in machine-generated text as compared to human ones.

  • Detecting computer-generated disinformation. [paper]

    Harald Stiff and Fredrik Johansson.

    Evaluate promising Transformer-based detection algorithms in a large variety of experiments involving both in-distribution and out-of-distribution test data, as well as evaluation on more realistic in-the-wild data.

  • On pushing DeepFake Tweet Detection capabilities to the limits. [paper]

    Margherita Gambini, Tiziano Fagni, Fabrizio Falchi and Maurizio Tesconi.

    Study and improve the performance of the stateof-the-art deepfake tweet detection methods over GPT-2 tweets and the detectors’ capabilities to generalize over tweets generated by GPT-3.

  • Cross-Domain Detection of GPT-2-Generated Technical Text. [paper]

    Juan Rodriguez, Todd Hay, David Gros, Zain Shamsi and Ravi Srinivasan.

    Find that RoBERTa-based detectors can be successfully adapted from one scientific discipline (physics) to another (biomedicine), requiring relatively small amounts of in-domain labeled data.

  • Automatic Detection of Machine Generated Texts: Need More Tokens. [paper]

    Paras Bhatt and Anthony Rios.

    Present a dataset for Russian language and conduct a set of learning experiments to build accurate machine-generated text detectors for both English and Russian languages.

  • Threat Scenarios and Best Practices to Detect Neural Fake News. [paper]

    Artidoro Pagnoni, Martin Graciarena and Yulia Tsvetkov.

    Provide an assessment of the current landscape of generated text detection and identify three primary threat scenarios. Establish the minimax strategies that minimize the worst case scenario for the detector.

  • Demystifying Neural Fake News via Linguistic Feature-Based Interpretation. [paper]

    Ankit Aich, Souvik Bhattacharya and Natalie Parde.

    Conduct a linguistically interpretative examination of the feature vulnerabilities exploited by neural fake news generators. Establish 21 stylistic, complexity, and psychological features for detection.

  • CoCo: Coherence-Enhanced Machine-Generated Text Detection Under Data Limitation With Contrastive Learning. [paper]

    Xiaoming Liu, Zhaohan Zhang, Yichen Wang, Yu Lan and Chao Shen.

    Present a coherence-based contrastive learning model named CoCo to detect the possible machine-generated text under low-resource scenario.

  • How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection. [paper]

    Biyang Guo, Xin Zhang, Ziyuan Wang, Minqi Jiang, Jinran Nie, Yuxuan Ding, Jianwei Yue and Yupeng Wu.

    Collect dataset the Human ChatGPT Comparison Corpus (HC3) and build three different detection systems to detect whether a certain text is generated by ChatGPT or humans.

  • DetectGPT: Zero-Shot Machine-Generated Text Detection using Probability Curvature. [paper]

    Eric Mitchell, Yoonho Lee, Alexander Khazatsky, Christopher D. Manning and Chelsea Finn.

    Propose DetectGPT, a zero-shot method for automated machine-generated text detection. To test if a passage came from a source model, DetectGPT compares the log probability of a candidate passage with the average log probability of several perturbations of the passage.

  • ChatGPT or Human? Detect and Explain. Explaining Decisions of Machine Learning Model for Detecting Short ChatGPT-generated Text. [paper]

    Sandra Mitrović, Davide Andreoletti and Omran Ayoub.

    Build Transformer-based model to discriminate between human-written and seemingly human text, focusing on short texts. Give some insights about ChatGPT writing style by SHAP explanations of the predictions.

  • AI vs. Human -- Differentiation Analysis of Scientific Content Generation. [paper]

    Yongqiang Ma, Jiawei Liu, Fan Yi, Qikai Cheng, Yong Huang, Wei Lu and Xiaozhong Liu.

    Investigate the detection methods from three perspectives: (1) feature-based detection method; (2) fine-tuned pre-trained detection model; (3) explainability of detection model. Further investigate the gap between AI-generated text and human-written text.

  • ChatGPT Generated Text Detection. [paper]

    Rexhep Shijaku and Ercan Canhasi.

    Present a classification model based on XGBoost for automatically detecting essays generated by ChatGPT.

  • Linguistic Markers of AI-Generated Text Versus Human-Generated Text: Evidence from Hotel Reviews and News Headlines. [paper]

    David M. Markowitz, Jeffrey Hancock and Jeremy Bailenson.

    Use ChatGPT to compare how it wrote hotel reviews and news headlines to human-generated counterparts across content, style, and structural features.

  • Mutation-Based Adversarial Attacks on Neural Text Detectors. [paper]

    Gongbo Liang, Jesus Guerrero and Izzat Alsmadi.

    Propose character- and word-based mutation operators for generating adversarial samples to attack state-of-the-art natural text detectors.

  • GAN-Based Unsupervised Learning Approach to Generate and Detect Fake News. [paper]

    Pranjal Bhardwaj, Krishna Yadav, Hind Alsharif and Rania Anwar Aboalela.

    Develop an unsupervised-based approach to generate fake news using autoencoder and GAN. Further describe the use of discriminators in detecting machine-generated fake news.

  • Combat AI With AI: Counteract Machine-Generated Fake Restaurant Reviews on Social Media. [paper]

    Alessandro Gambetti and Qiwei Han.

    Propose to leverage the high-quality elite restaurant reviews verified by Yelp to generate fake reviews from the OpenAI GPT review creator and ultimately fine-tune a GPT output detector to predict fake reviews that significantly outperform existing solutions.

  • Accurate Generated Text Detection Based on Deep Layer-wise Relevance Propagation. [paper]

    Mengjie Guo, Limin Liu, Meicheng Guo, Siyuan Liu and Zhiwei Xu.

    Propose a deep interpretable model to achieve accurate detection of the generated texts. In detail, a feature extraction method based on Layer-wise Relevance Propagation is proposed to improve the feature mining performance for detection of generated texts.

  • Stylometric Detection of AI-Generated Text in Twitter Timelines. [paper]

    Tharindu Kumarage, Joshua Garland, Amrita Bhattacharjee, Kirill Trapeznikov, Scott Ruston and Huan Liu.

    Propose two simple architectures to utilize three categories of stylometric features towards 1) discriminating between human-written and AI-generated tweets and 2) detecting if and when an AI starts to generate tweets in a given Twitter timeline.

  • ChatGPT or academic scientist? Distinguishing authorship with over 99% accuracy using off-the-shelf machine learning tools. [paper]

    Heather Desaire, Aleesa E. Chua, Madeline Isom, Romana Jarosova and David Hua.

    With a set of 20 features, build a model that assigned the author, as human or AI, at well over 99% accuracy, resulting in 20 times fewer misclassified documents compared to the field-leading approach.

  • Detection of AI-generated Essays in Writing Assessments. [paper]

    Duanli Yan, Michael Fauss, Jiangang Hao, Wenju Cui.

    Show how AI-generated essays are similar or different from human-written essays based on a set of typical prompts for a sample from a large-scale assessment. Introduce two classifiers that can detect AI-generated essays with a high accuracy of over 95%.

  • Comparing Abstractive Summaries Generated by ChatGPT to Real Summaries Through Blinded Reviewers and Text Classification Algorithms. [paper]

    Mayank Soni, Vincent Wade.

    Evaluate the performance of ChatGPT on Abstractive Summarization by the means of automated metrics and blinded human reviewers. Also build automatic text classifiers to detect ChatGPT generated summaries. Find that while text classification algorithms can distinguish between real and generated summaries, humans are unable to distinguish between real summaries and those produced by ChatGPT.

  • Distinguishing ChatGPT(-3.5, -4)-generated and human-written papers through Japanese stylometric analysis. [paper]

    Wataru Zaitsu and Mingzhe Jin.

    Perform multi-dimensional scaling (MDS) to confirm the distributions of 216 texts of three classes (human, GPT-3.5, GPT-4) focusing on the following stylometric features: (1) bigrams of parts-of-speech, (2) bigram of postpositional particle words, (3) positioning of commas, and (4) rate of function words.

  • Evaluating AIGC Detectors on Code Content. [paper]

    Jian Wang, Shangqing Liu, Xiaofei Xie and Yi Li.

    Present the first empirical study on evaluating existing AIGC detectors in the software domain. Create a comprehensive dataset including 492.5K samples comprising code-related content produced by ChatGPT and evaluate six AIGC detectors on this dataset.

  • Detection of Fake Generated Scientific Abstracts. [paper]

    Panagiotis C. Theocharopoulos, Panagiotis Anagnostou, Anastasia Tsoukala, Spiros V. Georgakopoulos, Sotiris K. Tasoulis and Vassilis P. Plagianakos.

    Utilize the GPT-3 model to generate scientific paper abstracts through Artificial Intelligence and explore various text representation methods when combined with Machine Learning models with the aim of identifying machine-written text.

  • ArguGPT: evaluating, understanding and identifying argumentative essays generated by GPT models. [paper]

    Yikang Liu, Ziyin Zhang, Wanyang Zhang, Shisen Yue, Xiaojing Zhao, Xinyuan Cheng, Yiwen Zhang and Hai Hu.

    Present ArguGPT, a balanced corpus of 4,038 argumentative essays generated by 7 GPT models in response to essay prompts from three sources: (1) in-class or homework exercises, (2) TOEFL and (3) GRE writing tasks. A RoBERTa fine-tuned with the training set of ArguGPT achieves above 90% accuracy in both essay- and sentence-level classification.

  • Improving Detection of ChatGPT-Generated Fake Science Using Real Publication Text: Introducing xFakeBibs a Supervised Learning Network Algorithm. [paper]

    Ahmed Abdeen Hamed.

    By means of a new algorithm, called xFakeBibs, show the significant difference between ChatGPT-generated fake publications and real publications. The algorithm predicted 98 of the 100 publications as fake, while 2 articles failed the test and were classified as real publications.

  • Differentiate ChatGPT-generated and Human-written Medical Texts. [paper]

    Wenxiong Liao, Zhengliang Liu, Haixing Dai, Shaochen Xu, Zihao Wu, Yiyang Zhang, Xiaoke Huang, Dajiang Zhu, Hongmin Cai, Tianming Liu and Xiang Li.

    Construct a suite of datasets containing medical texts written by human experts and generated by ChatGPT. Analyze the linguistic features of these two types of content and uncover differences in vocabulary, part-of-speech, dependency, sentiment, perplexity, etc. Finally, design and implement machine learning methods to detect medical text generated by ChatGPT.

  • CHEAT: A Large-scale Dataset for Detecting ChatGPT-writtEn AbsTracts. [paper]

    Peipeng Yu, Jiahan Chen, Xuan Feng and Zhihua Xia.

    Present a large-scale CHatGPT-writtEn AbsTract dataset (CHEAT) contains 35,304 synthetic abstracts, with Generation, Polish, and Mix as prominent representatives. Show that ChatGPT-written abstracts are detectable, while the detection difficulty increases with human involvement.

  • Origin Tracing and Detecting of LLMs. [paper]

    Linyang Li, Pengyu Wang, Ke Ren, Tianxiang Sun and Xipeng Qiu.

    First raise the concern of the origin tracing of LLMs and propose an effective method to trace and detect AI-generated contexts. Introduce a novel algorithm that leverages the contrastive features between LLMs and extracts model-wise features to trace the text origins.

  • ChatLog: Recording and Analyzing ChatGPT Across Time. [paper]

    Shangqing Tu, Chunyang Li, Jifan Yu, Xiaozhi Wang, Lei Hou and Juanzi Li.

    Collect a coarse-to-fine temporal dataset called ChatLog, consisting of two parts that update monthly and daily. In section 5, explore the application of ChatGPT unchanged features on ChatLog-Daily dataset. Show a preliminary use case: improving the ChatGPT detection model’s robustness.

  • Perception, performance, and detectability of conversational artificial intelligence across 32 university courses. [paper]

    Show AuthorsHazem Ibrahim, Fengyuan Liu, Rohail Asim, Balaraju Battu, Sidahmed Benabderrahmane, Bashar Alhafni, Wifag Adnan, Tuka Alhanai, Bedoor AlShebli, Riyadh Baghdadi, Jocelyn J. Bélanger, Elena Beretta, Kemal Celik, Moumena Chaqfeh, Mohammed F. Daqaq, Zaynab El Bernoussi, Daryl Fougnie, Borja Garcia de Soto, Alberto Gandolfi, Andras Gyorgy, Nizar Habash, J. Andrew Harris, Aaron Kaufman, Lefteris Kirousis, Korhan Kocak, Kangsan Lee, Seungah S. Lee, Samreen Malik, Michail Maniatakos, David Melcher, Azzam Mourad, Minsu Park, Mahmoud Rasras, Alicja Reuben, Dania Zantout, Nancy W. Gleason, Kinga Makovi, Talal Rahwan and Yasir Zaki

    Compare the performance of ChatGPT against students on 32 university-level courses. Current AI-text classifiers cannot reliably detect ChatGPT’s use in school work, due to their propensity to classify human-written answers as AI-generated, as well as the ease with which AI-generated text can be edited to evade detection.

  • GPT-Sentinel: Distinguishing Human and ChatGPT Generated Content. [paper]

    Yutian Chen, Hao Kang, Vivian Zhai, Liangze Li, Rita Singh and Bhiksha Raj.

    Collecte and release a pre-processed dataset named OpenGPTText, which consists of rephrased content generated using ChatGPT. Design, implement, and train two different models for text classification, using RoBERTa and T5, respectively. These models achieve remarkable results, with an accuracy of over 97% on the test dataset, as evaluated through various metrics.

  • Machine-Made Media: Monitoring the Mobilization of Machine-Generated Articles on Misinformation and Mainstream News Websites. [paper]

    Hans W. A. Hanley and Zakir Durumeric.

    Train a DeBERTa-based synthetic news detector and classify over 12.91 million articles from 3,074 misinformation and mainstream news websites. Find that between January 1, 2022 and April 1, 2023, the relative number of synthetic news articles increased by 79.4% on mainstream websites while increasing by 342% on misinformation sites.

  • Smaller Language Models are Better Black-box Machine-Generated Text Detectors. [paper]

    Fatemehsadat Mireshghallah, Justus Mattern, Sicun Gao, Reza Shokri and Taylor Berg-Kirkpatrick.

    Explore whether models other than the generator can be used to differentiate between machinegenerated and human-written text. Find that overall, smaller and partially-trained models are better universal text detectors. Interestingly, we find that whether the detector and generator were trained on the same data is not critically important to the detection success.

  • GPT Paternity Test: GPT Generated Text Detection with GPT Genetic Inheritance. [paper]

    Xiao Yu, Yuang Qi, Kejiang Chen, Guoqiang Chen, Xi Yang, Pengyuan Zhu, Weiming Zhang and Nenghai Yu.

    Introduce GPT Paternity Test (GPT-Pat), which reliably detects machine-generated text across varied datasets. Given a text under scrutiny, we leverage ChatGPT to generate a corresponding question and provide a re-answer to the question. By comparing the similarity between the original text and the generated re-answered text, it can be determined whether the text is machine-generated.

  • G3Detector: General GPT-Generated Text Detector. [paper]

    Haolan Zhan, Xuanli He, Qiongkai Xu, Yuxiang Wu and Pontus Stenetorp.

    Introduce an unpretentious yet potent detection approach proficient in identifying synthetic text across a wide array of fields. The detector demonstrates outstanding performance uniformly across various model architectures and decoding strategies. It also possesses the capability to identify text generated utilizing a potent detectionevasion technique.

  • Deepfake Text Detection in the Wild. [paper]

    Yafu Li, Qintong Li, Leyang Cui, Wei Bi, Longyue Wang, Linyi Yang, Shuming Shi and Yue Zhang.

    Construct a wild testbed for deepfake text detection, by gathering texts from various writing tasks and deepfake texts generated by different LLMs. Find that supervised PLM-based methods were the best-performing detection methods across all testbeds. Out-of-distribution posed an even greater challenge for a detector to be employed in realistic application scenarios.

  • DetectLLM: Leveraging Log Rank Information for Zero-Shot Detection of Machine-Generated Text. [paper]

    Jinyan Su, Terry Yue Zhuo, Di Wang and Preslav Nakov.

    Introduce two novel zero-shot methods for detecting machinegenerated text by leveraging the log rank information. One is called DetectLLM-LRR, which is fast and efficient, and the other is called DetectLLM-NPR, which is more accurate, but slower due to the need for perturbations.

  • LLMDet: A Third Party Large Language Models Generated Text Detection Tool. [paper]

    Kangxi Wu, Liang Pang, Huawei Shen, Xueqi Cheng and Tat-Seng Chua.

    Propose an efficient, secure, and scalable detection tool called LLMDet, which calculates the proxy perplexity of text by utilizing the prior information of the model’s next-token probabilities, obtained through pre-training. Subsequently, use the self-watermarking information of the model, as measured by proxy perplexity, to detect the source of the text.

  • Ghostbuster: Detecting Text Ghostwritten by Large Language Models. [paper]

    Vivek Verma, Eve Fleisig, Nicholas Tomlin and Dan Klein.

    Introduce Ghostbuster, works by passing documents through a series of weaker language models and running a structured search over possible combinations of their features, then training a classifier on the selected features to determine if the target document was AI-generated. Release three new datasets of human and AI-generated text as detection benchmarks that cover multiple domains and task setups.

  • HowkGPT: Investigating the Detection of ChatGPT-generated University Student Homework through Context-Aware Perplexity Analysis. [paper]

    Christoforos Vasilatos, Manaar Alam, Talal Rahwan, Yasir Zaki and Michail Maniatakos.

    Propose a novel multi-level approach to detect AIgenerated text focusing on university student homework. Utilize metadata categorization from an academic dataset to enhance the perplexity metric used to detect whether a given assignment has been studentauthored or AI-generated.

  • DNA-GPT: Divergent N-Gram Analysis for Training-Free Detection of GPT-Generated Text. [paper]

    Xianjun Yang, Wei Cheng, Linda Petzold, William Yang Wang and Haifeng Chen.

    Given a text, we first truncate it in the middle and then use only the preceding portion as input to the LLMs to regenerate the new remaining parts. By analyzing the differences between the original and new remaining parts through N-gram analysis in black-box or probability divergence in white-box, we can clearly illustrate significant discrepancies between machine-generated and human-written text.

  • Efficient Detection of LLM-generated Texts with a Bayesian Surrogate Model. [paper]

    Zhijie Deng, Hongcheng Gao, Yibo Miao and Hao Zhang.

    DetectGPT suffers from significant inefficiency issues, as detecting a single candidate requires scoring hundreds of its perturbations with the source LLM. This paper aims to bridge this gap. Technically, propose to incorporate a Bayesian surrogate model, which allows us to select typical samples based on Bayesian uncertainty and interpolate scores from typical samples to other ones, to improve query efficiency.

  • Distinguishing Human Generated Text From ChatGPT Generated Text Using Machine Learning. [paper]

    Niful Islam, Debopom Sutradhar, Humaira Noor, Jarin Tasnim Raya, Monowara Tabassum Maisha and Dewan Md Farid.

    Present a machine learning-based solution that can identify the ChatGPT delivered text from the human written text along with the comparative analysis of a total of 11 machine learning and deep learning algorithms in the classification process.

  • Multiscale Positive-Unlabeled Detection of AI-Generated Texts. [paper]

    Yuchuan Tian, Hanting Chen, Xutao Wang, Zheyuan Bai, Qinghua Zhang, Ruifeng Li, Chao Xu and Yunhe Wang.

    Mainstream detectors are formulated without considering the factor of corpus length: shorter corpuses are harder to detect compared with longer ones for shortage of informative features. In this paper, a Multiscale Positive-Unlabeled (MPU) training framework is proposed to address the challenge of multiscale text detection.

  • Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts. [paper]

    Eduard Tulchinskii, Kristian Kuznetsov, Laida Kushnareva, Daniil Cherniavskii, Serguei Barannikov, Irina Piontkovskaya, Sergey Nikolenko and Evgeny Burnaev.

    Propose an invariant of human texts, namely the intrinsic dimensionality of the manifold underlying the set of embeddings of a given text sample. We show a clear statistical separation between human-generated and AI-generated distributions. This property allows us to build a score-based artificial text detector, which outperforming SOTA detectors in model-agnostic and cross-domain scenarios by a significant margin.

  • Distinguishing academic science writing from humans or ChatGPT with over 99% accuracy using off-the-shelf machine learning tools. [paper]

    Heather Desaire, Aleesa E. Chua, Madeline Isom, Romana Jarosova and David Hua.

    Report a method for discriminating text generated by ChatGPT from (human) academic scientists, relying on prevalent and accessible supervised classification methods. The approach uses new features for discriminating (these) humans from AI.

  • Long-form analogies generated by chatGPT lack human-like psycholinguistic properties. [paper]

    S. M. Seals and Valerie L. Shalin.

    Apply psycholinguistic methods to evaluate individual sentences from long-form analogies about biochemical concepts. Perform a supervised classification analysis using 78 features extracted from Coh-metrix that analyze text cohesion, language, and readability. Results illustrate high performance for classifying student-generated and chatGPT-generated analogies.

  • Check Me If You Can: Detecting ChatGPT-Generated Academic Writing using CheckGPT. [paper]

    Zeyan Liu, Zijun Yao, Fengjun Li and Bo Luo.

    Present GPABenchmark, a benchmarking dataset of 600,000 samples of human-written, GPT-written, GPT-completed, and GPT-polished abstracts of research papers in CS, physics, and humanities and social sciences (HSS). Present CheckGPT, a novel LLM-content detector consisting of a general representation module and an attentive-BiLSTM classification module, which is accurate, transferable, and interpretable.

  • Stanford CS224N Custom Projects

    • Check Me If You Can: Detecting ChatGPT-Generated Academic Writing using CheckGPT. [paper]

      Jeffrey Heo and Simon Kim.

    • Looking Under the Hood of DetectGPT. [paper]

      Maximilian Du, Ryan Lian and Kaien Yang.

    • GPTNo: A Deep Learning LLM to Beat the Turing Test. [paper]

      Sri Jaladi, William Li and Abhinav Sinha.

    • Human Writing is as Uniform as Machine Writing. [paper]

      Ryan Tan, Raghav Garg and Jacob Stavrianos.

    • DetectChatGPT: Black-Box Zero-Shot Detection of LLM-Generated Text. [paper]

      Julia Park and Armaan Rashid.

  • La détection de textes générés par des modèles de langue: une tâche complexe? Une étude sur des textes académiques. [paper]

    Vijini Liyanage and Davide Buscaldi.

  • Towards a Robust Detection of Language Model Generated Text: Is ChatGPT that Easy to Detect? [paper]

    Wissam Antoun, Virginie Mouilleron, Benoît Sagot and Djamé Seddah.

    Proposes a methodology for developing and evaluating ChatGPT detectors for French text, with a focus on investigating their robustness on outof-domain data and against common attack schemes. The proposed method involves translating an English dataset into French and training a classifier on the translated data.

  • Implementing BERT and fine-tuned RobertA to detect AI generated news by ChatGPT. [paper]

    Zecong Wang, Jiaxi Cheng, Chen Cui and Chenhao Yu.

    Demonstrate that the BERT and RobertA models with fine-tuning had the best success in detecting AI generated news. With a score of 98%, tweaked RobertA in particular showed excellent precision.

  • Beyond Black Box AI-Generated Plagiarism Detection: From Sentence to Document Level. [paper]

    Mujahid Ali Quidwai, Chunhui Li and Parijat Dube.

    Propose a novel approach offering quantifiable metrics at both sentence. Employs a multi-faceted approach, generating multiple paraphrased versions of a given question and inputting them into the LLM to generate answers. By using a contrastive loss function based on cosine similarity, match generated sentences with those from the student’s response.

  • Matching Pairs: Attributing Fine-Tuned Models to their Pre-Trained Large Language Models. [paper]

    Myles Foley, Ambrish Rawat, Taesung Lee, Yufang Hou, Gabriele Picco and Giulio Zizzo.

    We need a method to investigate how a model was trained or a piece of text was generated and what their pre-trained base model was. Address this open problem by tracing back the origin of a given fine-tuned LLM to its corresponding pre-trained base model. Consider different knowledge levels and attribution strategies, and find that we can correctly trace back 8 out of the 10 fine tuned models with our best method.

  • Detecting Artificially Generated Academic Text: The Importance of Mimicking Human Utilization of Large Language Models. [paper]

    Vijini Liyanage and Davide Buscaldi.

    Examine the detectability of the generated text using DetectGPT and GLTR, and utilize state-of-the-art classification models like SciBERT, RoBERTa, DEBERTa, XLNet, and ELECTRA. Experiments show that the generated text is difficult to detect using existing models when created using a LLM fine-tuned on the remainder of a paper.

  • Testing of Detection Tools for AI-Generated Text. [paper]

    Debora Weber-Wulff, Alla Anohina-Naumeca, Sonja Bjelobaba, Tomáš Foltýnek, Jean Guerrero-Dib, Olumide Popoola, Petr Šigut and Lorna Waddington.

    The study seeks to answer research questions about whether existing detection tools can reliably differentiate between human-written text and ChatGPT-generated text, and whether machine translation and content obfuscation techniques affect the detection of AIgenerated text. The research covers 12 publicly available tools and two commercial systems (Turnitin and PlagiarismCheck) that are widely used in the academic setting.

  • Discriminating Human-authored from ChatGPT-Generated Code Via Discernable Feature Analysis. [paper]

    Li Ke, Hong Sheng, Fu Cai, Zhang Yunhe and Liu Ming.

    Propose a discriminative feature set yielding high accuracy in differentiating ChatGPT-generated code from human-authored code in binary classification tasks; Devise methods for generating extensive ChatGPT-generated codes; Introduce a dataset cleansing strategy.

  • RADAR: Robust AI-Text Detection via Adversarial Learning. [paper]

    Xiaomeng Hu, Pin-Yu Chen and Tsung-Yi Ho.

    Existing works show that current AI-text detectors are not robust to LLM-based paraphrasing, this paper aims to bridge this gap by proposing a new framework called RADAR. RADAR is based on adversarial training of a paraphraser and a detector. The paraphraser’s goal is to generate realistic contents to evade AI-text detection. RADAR uses the feedback from the detector to update the paraphraser, and vice versa.

  • Detecting LLM-Generated Text in Computing Education: A Comparative Study for ChatGPT Cases. [paper]

    Michael Sheinman Orenstrakh, Oscar Karnalim, Carlos Anibal Suarez and Michael Liut.

    Collect 124 submissions from computer science students before the creation of ChatGPT, then generate 40 ChatGPT submissions. Use this data to evaluate eight publicly-available LLM-generated text detectors through the measures of accuracy, false positives, and resilience.

  • Is ChatGPT Involved in Texts? Measure the Polish Ratio to Detect ChatGPT-Generated Text. [paper]

    Lingyi Yang, Feng Jiang and Haizhou Li.

    Introduce a novel dataset termed HPPT (ChatGPT-polished academic abstracts), facilitating the construction of more robust detectors. Propose the "Polish Ratio" method, an innovative measure of ChatGPT’s involvement in text generation based on editing distance. It provides a mechanism to measure the degree of human originality in the resulting text.

  • OUTFOX: LLM-generated Essay Detection through In-context Learning with Adversarially Generated Examples. [paper]

    Ryuto Koike, Masahiro Kaneko and Naoaki Okazaki.

    Propose OUTFOX, a novel framework that improves the robustness of LLM-generated-text detectors by allowing both the detector and the attacker to consider each other’s output and apply this to the domain of student essays.

  • The Looming Threat of Fake and LLM-generated LinkedIn Profiles: Challenges and Opportunities for Detection and Prevention. [paper]

    Navid Ayoobi, Sadat Shahriar and Arjun Mukherjee.

    Present a novel method for detecting fake and Large Language Model (LLM)-generated profiles in the LinkedIn Online Social Network immediately upon registration and before establishing connections.

  • The Imitation Game: Detecting Human and AI-Generated Texts in the Era of Large Language Models. [paper]

    Kadhim Hayawi, Sakib Shahriar and Sujith Samuel Mathew.

    Introduce a novel dataset of human-written and LLM-generated texts in different genres: essays, stories, poetry, and Python code. Employ several machine learning models to classify the texts.

  • Towards Automatic Boundary Detection for Human-AI Collaborative Hybrid Essay in Education. [paper]

    Zijie Zeng, Lele Sha, Yuheng Li, Kaixun Yang, Dragan Gašević and Guanliang Chen.

    Construct a hybrid essay dataset. Propose a two-step detection approach where we (1) separate AI-generated content from human-written content during the encoder training process; and (2) calculate the distances between every two adjacent prototypes and assume that the boundaries exist between the two adjacent prototypes that have the furthest distance from each other.

  • Fighting Fire with Fire: Can ChatGPT Detect AI-generated Text? [paper]

    Amrita Bhattacharjee and Huan Liu.

    Empirically investigate if ChatGPT is symmetrically effective in detecting AI-generated or human-written text. Provide insight on how ChatGPT and similar LLMs may be leveraged in automated detection pipelines by simply focusing on solving a specific aspect of the problem and deriving the rest from that solution.

  • A Deep Fusion Model for Human vs. Machine-Generated Essay Classification. [paper]

    Roberto Corizzo and Sebastian Leal-Arenas.

    Propose a deep neural network-based model that combines text modeling and linguistic features via data fusion, with the aim to provide robust classification capabilities. Experiments show that it is possible to accurately perform essay classification, and that our model is competitive, even when compared to popular feature-based and neural network-based approaches.

  • MFD: Multi-Feature Detection of LLM-Generated Text. [paper]

    Zhendong Wu and Hui Xiang.

    Introduce Multi-Feature Detection (MFD), a new zero-shot method. MFD comprehensively considers log-likelihood, log-rank, entropy, and LLM-Deviation. LLM-Deviation is a new statistical feature proposed in this paper and has a clear distribution difference between texts generated by LLMs and those written by humans. Experiments show MFD is more effective than the existing zero-shot method.

  • Classification of Human- and AI-Generated Texts: Investigating Features for ChatGPT. [paper]

    Lorenz Mindner, Tim Schlippe and Kristina Schaaff.

    Explore traditional and new features to (1) detect text generated by AI from scratch and (2) text rephrased by AI. Results show that the new features substantially help to improve the performance of many classifiers.

  • Separating the Human Touch from AI-Generated Text using Higher Criticism: An Information-Theoretic Approach. [paper]

    Alon Kipnis.

    Propose a method to determine whether a given article was entirely written by a generative language model. The process involves many perplexity tests for the origin of individual sentences or other text atoms, combining these multiple tests using Higher Criticism (HC).

  • 人工智能生成语言与人类语言对比研究——以ChatGPT为例. [paper]

    朱君辉, 王梦焰, 杨尔弘, 聂锦燃, 王誉杰, 岳岩 and 杨麟儿.

    Explore the differences between the language used in human-generated responses and responses generated by ChatGPT. Extract and compute the distribution of 159 language features in real human text and ChatGPT-generated text. Employ three machine learning algo-rithms: Random Forest, Logistic Regression, and SVM. The result reveals that the two texts differ significantly in three dimensions: descriptive features, word commonness, and word diversity.

  • HC3 Plus: A Semantic-Invariant Human ChatGPT Comparison Corpus. [paper]

    Zhenpeng Su, Xing Wu, Wei Zhou, Guangyuan Ma and Songlin Hu.

    Introduce a more extensive and comprehensive dataset that considers more types of tasks than previous work, including semantic-invariant tasks. Further instruct fine-tuning Tk-instruct and build a more powerful detection system. Experimental results show that the proposed detector outperforms the previous state-of-the-art RoBERTa-based detector.

  • J-Guard: Journalism Guided Adversarially Robust Detection of AI-generated News. [paper]

    Tharindu Kumarage, Amrita Bhattacharjee, Djordje Padejski, Kristy Roschke, Dan Gillmor, Scott Ruston, Huan Liu and Joshua Garland.

    Develop a framework, J-Guard, capable of steering existing supervised AI text detectors for detecting AI-generated news while boosting adversarial robustness. By incorporating stylistic cues inspired by the unique journalistic attributes, J-Guard effectively distinguishes between real-world journalism and AIgenerated news articles.

  • Overview of AuTexTification at IberLEF 2023: Detection and Attribution of Machine-Generated Text in Multiple Domains. [paper]

    Areg Mikael Sarvazyan, José Ángel González, Marc Franco-Salvador, Francisco Rangel, Berta Chulvi and Paolo Rosso.

    This paper presents the overview of the AuTexTification shared task as part of the IberLEF 2023 Workshop. AuTexTification consists of two subtasks: for Subtask 1, participants had to determine whether a text is human-authored or has been generated by a large language model. For Subtask 2, participants had to attribute a machine-generated text to one of six different text generation models. In this overview, present the AuTexTification dataset and task, the submitted participating systems, and the results.

  • ConDA: Contrastive Domain Adaptation for AI-generated Text Detection. [paper]

    Amrita Bhattacharjee, Tharindu Kumarage, Raha Moraffah and Huan Liu.

    Address the problem of AIgenerated text detection in the absence of labeled target data. Propose a contrastive domain adaptation framework that leverages the power of both unsupervised domain adaptation and selfsupervised representation learning, in order to tackle the task of AI-generated text detection.

  • Supervised Machine-Generated Text Detectors: Family and Scale Matters. [paper]

    Areg Mikael Sarvazyan, José Ángel González, Paolo Rosso and Marc Franco-Salvador.

    This work studies the generalization capabilities of supervised Machine-Generated Text (MGT) detectors across model families and parameter scales of text generation models. In addition, explore the feasibility of identifying the family and scale of the generator behind an MGT, instead of attributing the text to a particular language model.

  • Use prompt to differentiate text generated by ChatGPT and humans. [paper]

    Ruopeng An, Yuyi Yang, Fan Yang and Shanshan Wang.

    Introduce a method that uses both prompts and essays to differentiate between machine-generated and human-written text. Validate its effectiveness in distinguishing between machine-generated and human-written essays and shows that it outperforms existing approaches based solely on text passages.

  • A Statistical Turing Test for Generative Models. [paper]

    Hayden Helm, Carey E. Priebe and Weiwei Yang.

    Provide a framework in the language of statistical pattern recognition that quantifies the difference between the distributions of human and machinegenerated content conditioned on an evaluation context. Describe current methods in the context of the framework and demonstrate how to use the framework to evaluate the progression of generative models towards human-like capabilities, among many axes of analysis.

  • TopRoBERTa: Topology-Aware Authorship Attribution of Deepfake Texts. [paper]

    Adaku Uchendu, Thai Le and Dongwon Lee.

    Propose TopRoBERTa to improve existing AA solutions by capturing more linguistic patterns in deepfake texts by including a Topological Data Analysis (TDA) layer in the RoBERTa model. Use RoBERTa to capture contextual representations (i.e., semantic and syntactic linguistic features), while using TDA to capture the shape and structure of data (i.e., linguistic structures). Finally, TopRoBERTa, outperforms the vanilla RoBERTa in 2/3 datasets, achieving up to 7% increase in Macro F1 score.

  • From Text to Source: Results in Detecting Large Language Model-Generated Content. [paper]

    Wissam Antoun, Benoît Sagot and Djamé Seddah.

    Investigate "Cross-Model Detection," evaluating whether a classifier trained to distinguish between source LLM-generated and humanwritten text can also detect text from a target LLM without further training. The study comprehensively explores various LLM sizes and families, and assesses the impact of conversational fine-tuning techniques on classifier generalization. The research also delves into Model Attribution, encompassing source model identification, model family classification, and model size classification.

  • Can LLM-Generated Misinformation Be Detected? [paper]

    Canyu Chen and Kai Shu.

    Build a taxonomy of LLM-generated misinformation. Categorize and validate the potential real-world methods for generating misinformation with LLMs. Discover that LLM-generated misinformation can be harder to detect for humans and detectors compared to human-written misinformation with the same semantics.

  • A Framework for Detecting AI-Generated Text in Research Publications. [paper]

    Paria Sarzaeim, Arya Doshi and Qusay Mahmoud.

    Present a framework for AI-generated text. The prototype implementation of the proposed approach is to train a model using predefined datasets and deploy this model on a cloud-based service to predict whether a text was created by a human or AI. This approach is specifically focused on assessing the accuracy of scientific writings and research papers rather than general text.

  • Supervised Machine Generated Text Detection Using LLM Encoders In Various Data Resource Scenarios. [paper]

    Marc Capobianco, Matthew Reynolds, Charles Phelan, Krish Shah-Nathwani and Duong Luong.

    Explore the effectiveness of BERT and RoBERTa-based machine generated text detection in a supervised setting. Created several models for both BERT and RoBERTa, trained with 5%, 10%, 15%, 20%, and 100% of the dataset. Conduct these experiments with frozen and unfrozen parameter variations.

  • Machine-Generated Text Detection and Attribution. [paper]

    Sarvazyan and Areg Mikael.

    (i) study the MGT detection and attribution models that achieve high performance under various settings, (ii) explore their generalization capabilities to different scenarios, analyzing their applicability to unseen types of data, and (iii) successfully organize an evaluation campaign for MGT detection and attribution.

  • Automated Text Identification Shared Task (AuTexTification)

    • Automated Text Identification Using CNN and Training Dynamics. [paper]

      Claudiu Creanga and Liviu Petrisor Dinu.

    • Do Origin and Facts Identify Automatically Generated Text? [paper]

      Judita Preiss and Monica Lestari Paramita.

    • Syntax-based Contrastive Learning for Automated Text Identification. [paper]

      Hongyan Wu, Nankai Lin and Shengyi Jiang.

    • A Method for Identifying Bot Generated Text : Notebook for IberLEF 2023. [paper]

      Maryam Elamine, Asma Mekki and Lamia Hadrich Belguith.

    • Artificial Intelligence-Based Text Classification: Separating Human Writing from Computer Generated Writing. [paper]

      Luis Enrique Morales-Márquez, Erick Barrios-González and David Eduardo Pinto-Avendaño.

    • SINAI at AuTexTification in IberLEF 2023: Combining All Layer Embeddings for Automatically Generated Texts. [paper]

      César Espin-Riofrio, Jenny Ortiz-Zambrano and Arturo Montejo-Ráez.

    • I’ve Seen Things You Machines Wouldn’t Believe: Measuring Content Predictability to Identify Automatically-Generated Text. [paper]

      Piotr Przybyła, Nicolau Duran-Silva and Santiago Egea-Gómez.

    • Detecting Generated Text and Attributing Language Model Source with Fine-tuned Models and Semantic Understanding. [paper]

      Margherita Gambini, Marco Avvenuti, Fabrizio Falchi, Maurizio Tesconi and Tiziano Fagni.

    • AI-Writing Detection Using an Ensemble of Transformers and Stylometric Features. [paper]

      George Mikros, Athanasios Koursaris, Dimitrios Bilianos and George Markopoulos.

    • Exploring Text Representations for Detecting Automatically Generated Text. [paper]

      Zayra Villegas-Trejo, Helena Gómez-Adorno and Sergio-Luis Ojeda-Trueba.

    • GPT-2 versus GPT-3 and Bloom: LLMs for LLMs Generative Text Detection. [paper]

      Fernando Aguilar-Canto, Marco Cardoso-Moreno, Diana Jiménez and Hiram Calvo.

    • Taming the Turing Test: Exploring Machine Learning Approaches to Discriminate Human vs. AI-Generated Texts. [paper]

      Alberto Fernández-Hernández, Juan Luis Arboledas-Márquez, Julián Ariza-Merino and Salud María Jiménez-Zafra.

    • Team GPLSI at AuTexTification Shared Task: Determining the Authorship of a Text. [paper]

      Iván Martínez-Murillo, Robiert Sepúlveda-Torres, Estela Saquete, Elena LLoret and Manuel Palomar.

    • Generative AI Text Classification using Ensemble LLM Approaches. [paper]

      Harika Abburi, Michael Suesserman, Nirmala Pudota, Balaji Veeramani, Edward Bowen and Sanmitra Bhattacharya.

    • Automated Text Identification: Multilingual Transformer-based Models Approach. [paper]

      German Gritsay, Andrey Grabovoy, Aleksandr Kildyakov and Yury Chekhovich.

    • IberLEF 2023 AuTexTification: Automated Text Identification Shared Task – Team OD-21. [paper]

      Rinaldo Gagiano, Haytham Fayek, Maria Myung-Hee Kim, Jennifer Biggs and Xiuzhen Zhang.

    • Using Linguistic Knowledge for Automated Text Identification. [paper]

      Lara Alonso Simón, José Antonio Gonzalo Gimeno, Ana María Fernández-Pampillón Cesteros, Marianela Fernández Trinidad and María Victoria Escandell Vidal.

    • Univ. of Hildesheim at AuTexTification 2023: Detection of Automatically Generated Texts. [paper]

      Tatjana Scheibe and Thomas Mandl.

    • UPB at IberLEF-2023 AuTexTification: Detection of Machine-Generated Text using Transformer Ensembles. [paper]

      Andrei-Alexandru Preda, Dumitru-Clementin Cercel, Traian Rebedea and Costin-Gabriel Chiru.

    • Exploring the Distinction: Investigating the Recognition of Automatic Text Generation Systems and Differentiating Human Text from Language Models. [paper]

      Manex Agirrezabal.

  • Looking for Traces of Textual Deepfakes in Bulgarian on Social Media. [paper]

    Irina Temnikova, Iva Marinova, Silvia Gargova, Ruslana Margova and Ivan Koychev.

    Create a new Bulgarian-language dataset with real social media messages and those generated by two language models. Conclude that combining LM text detection with fact-checking is the most appropriate method for this task, and that identifying Bulgarian textual deepfakes is indeed possible.

  • On the Generalization of Training-based ChatGPT Detection Methods. [paper]

    Han Xu, Jie Ren, Pengfei He, Shenglai Zeng, Yingqian Cui, Amy Liu, Hui Liu and Jiliang Tang.

    Aim to have a comprehensive investigation on existing methods’ generalization behaviors under distribution shift caused by a wide range of factors, including prompts, text lengths, topics, and language tasks. Collect a new dataset with human and ChatGPT texts, and then conduct extensive studies on the collected dataset.

  • Counter Turing Test CT^2: AI-Generated Text Detection is Not as Easy as You May Think -- Introducing AI Detectability Index. [paper]

    Megha Chakraborty, S.M Towhidul Islam Tonmoy, S M Mehedi Zaman, Krish Sharma, Niyar R Barman, Chandan Gupta, Shreya Gautam, Tanay Kumar, Vinija Jain, Aman Chadha, Amit P. Sheth and Amitava Das.

    Introduce the Counter Turing Test (CT2), a benchmark consisting of techniques aiming to offer a comprehensive evaluation of the robustness of existing AGTD techniques. Establish a quantifiable spectrum facilitating the evaluation and ranking of LLMs according to their detectability levels, we propose the AI Detectability Index (ADI).

  • Fast-DetectGPT: Efficient Zero-Shot Detection of Machine-Generated Text via Conditional Probability Curvature. [paper]

    Guangsheng Bao, Yanbin Zhao, Zhiyang Teng, Linyi Yang and Yue Zhang.

    Present Fast-DetectGPT, an optimized zero-shot detector, which substitutes DetectGPT’s perturbation step with a more efficient sampling step. The evaluations on various datasets, source models, and test conditions indicate that Fast-DetectGPT not only outperforms DetectGPT in both the whitebox and black-box settings but also accelerates the detection process by a factor of 340.

  • Zero-Shot Detection of Machine-Generated Codes. [paper]

    Xianjun Yang, Kexun Zhang, Haifeng Chen, Linda Petzold, William Yang Wang and Wei Cheng.

    Propose a training-free approach for the detection of LLMs-generated codes, mitigating the risks associated with their indiscriminate usage. Modify the previous zero-shot text detection method, DetectGPT by utilizing a surrogate white-box model to estimate the probability of the rightmost tokens, allowing us to identify code snippets generated by language models.

  • How Reliable Are AI-Generated-Text Detectors? An Assessment Framework Using Evasive Soft Prompts. [paper]

    Tharindu Kumarage, Paras Sheth, Raha Moraffah, Joshua Garland and Huan Liu.

    Suggest a universal evasive prompt, a novel type of soft prompt, which guides PLMs in producing "human-like" text that can mislead the detectors. The novel universal evasive prompt is achieved in two steps: First, create an evasive soft prompt tailored to a specific PLM through prompt tuning; and then, leverage the transferability of soft prompts to transfer the learned evasive soft prompt from one PLM to another.

  • On the Zero-Shot Generalization of Machine-Generated Text Detectors. [paper]

    Xiao Pu, Jingyu Zhang, Xiaochuang Han, Yulia Tsvetkov and Tianxing He.

    Collect generation data from a wide range of LLMs, and train neural detectors on data from each generator and test its performance on heldout generators. Demonstrate that robust detectors can be built on an ensemble of training data from medium-sized models.

  • GPT-who: An Information Density-based Machine-Generated Text Detector. [paper]

    Saranya Venkatraman, Adaku Uchendu and Dongwon Lee.

    Propose GPT-who, the first psycholinguistically-aware multi-class domainagnostic statistical-based detector. This detector employs UID-based features to model the unique statistical signature of each LLM and human author for accurate authorship attribution.

  • SeqXGPT: Sentence-Level AI-Generated Text Detection. [paper]

    Pengyu Wang, Linyang Li, Ke Ren, Botian Jiang, Dong Zhang and Xipeng Qiu.

    Propose Sequence X (Check) GPT, a novel method that utilizes log probability lists from white-box LLMs as features for sentence-level AIGT detection. These features are composed like waves in speech processing and cannot be studied by LLMs. Build SeqXGPT based on convolution and self-attention networks.

  • Raidar: geneRative AI Detection viA Rewriting. [paper]

    Chengzhi Mao, Carl Vondrick, Hao Wang and Junfeng Yang.

    Find that large language models (LLMs) are more likely to modify human-written text than AI-generated text when tasked with rewriting. Introduce a method to detect AI-generated content by prompting LLMs to rewrite text and calculating the editing distance of the output.

  • Spotting LLMs With Binoculars: Zero-Shot Detection of Machine-Generated Text. [paper]

    Abhimanyu Hans, Avi Schwarzschild, Valeriia Cherepanova, Hamid Kazemi, Aniruddha Saha, Micah Goldblum, Jonas Geiping and Tom Goldstein.

    Propose a novel LLM detector that only requires simple calculations using pre-trained LLMs. The method, called Binoculars, achieves state-of-the-art accuracy without any training data.

  • Human-in-the-loop Detection of AI-generated Text via Grammatical Patterns. [paper]

    Johan Lokna, Mislav Balunovic and Martin Vechev.

    Focus on the problem of detecting AI-generated text in a domain where a training dataset of human-written samples is readily available. Our key insight is to learn interpretable grammatical patterns that are highly indicative of human or AI written text. The most useful of these patterns can then be given to humans as part of a human-in-the-loop approach.

  • Few-Shot Detection of Machine-Generated Text using Style Representations. [paper]

    Rafael Rivera Soto, Kailin Koch, Aleem Khan, Barry Chen, Marcus Bishop and Nicholas Andrews.

    Pursue a approach not relying on samples from language models of concern at training time. Propose to leverage representations of writing style estimated from human-authored text. Find that features effective at distinguishing among human authors are also effective at distinguishing human from machine authors.

  • Metric Learning for Detection of Large Language Model Generated Texts. [paper]

    Linh Le, Shashank Hebbar and My Nguyen.

    Present a new paradigm of metric-based detection for LLM-generated texts that is able to balance among computational costs, accessibility, and performances. Specifically, the detection is performed through evaluating the similarity between a given text to an equivalent example generated by LLMs and through that determining the former's origination.

  • Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean Discrepancy. [paper]

    Shuhai Zhang, Feng Liu, Jiahao Yang, Yifan Yang, Changsheng Li, Bo Han and Mingkui Tan.

    Propose a novel multi-population aware optimization method for MMD called MMD-MP, which can avoid variance increases and thus improve the stability to measure the distributional discrepancy. Relying on MMD-MP, develop two methods for paragraph-based and sentence-based detection, respectively.

  • Using AI-based detectors to control AI-assisted plagiarism in ESL writing: “The Terminator Versus the Machines”. [paper]

    Karim Ibrahim.

    To address the challenge of AI-assisted plagiarism in ESL contexts, the present cross-disciplinary descriptive study examined the potential of two RoBERTa-based classifiers to control AI-assisted plagiarism on a dataset of 240 human-written and ChatGPT-generated essays. Data analysis revealed that both platforms could identify AI-generated texts, but their detection accuracy was inconsistent across the dataset.

  • An Ensemble-Based Approach for Generative Language Model Attribution. [paper]

    Harika Abburi, Michael Suesserman, Nirmala Pudota, Balaji Veeramani, Edward Bowen and Sanmitra Bhattacharya.

    Propose an ensemble neural model that generates probabilities from multiple pre-trained LLMs, which are then used as features for a traditional machine learning classifier.

  • DetectGPT-SC: Improving Detection of Text Generated by Large Language Models through Self-Consistency with Masked Predictions. [paper]

    Rongsheng Wang, Qi Li and Sihong Xie.

    Find that large language models such as ChatGPT exhibit strong self-consistency in text generation and continuation. Propose DetectGPT-SC for AI-generated texts detection based on self-consistency with masked predictions to determine whether a text is generated by LLMs.

  • Do Stochastic Parrots have Feelings Too? Improving Neural Detection of Synthetic Text via Emotion Recognition. [paper]

    Alan Cowap, Yvette Graham and Jennifer Foster.

    Hypothesize that pretrained language models (PLMs) have an affective deficit because of lacking the kind of emotional coherence present in human-authored text. Develop an emotionally aware detector by fine-tuning a PLM on emotion.

  • How well can machine-generated texts be identified and can language models be trained to avoid identification? [paper]

    Sinclair Schneider, Florian Steuber, Joao A. G. Schneider and Gabi Dreo Rodosek.

    Highlight the significance of sampling techniques and the essential role of employing advanced models like transformers to detect generated texts accurately. Demonstrate how a malicious actor could adapt generative models to evade a detector if accessible.

  • An Ensemble Method Based on the Combination of Transformers with Convolutional Neural Networks to Detect Artificially Generated Text. [paper]

    Vijini Liyanage and Davide Buscaldi.

    Present some classification models constructed by ensembling transformer models such as SciBERT, DeBERTa and XLNet, with Convolutional Neural Networks (CNNs). The experiments demonstrate that the considered ensemble architectures surpass the performance of the individual transformer models for classification.

  • Detection of news written by the ChatGPT through authorship attribution performed by a Bidirectional LSTM model. [paper]

    Amanda Ferrari Iaquinta and Gustavo Voltani von Atzingen.

    A dataset containing equal amounts of human and ChatGPT written news was assembled and different natural processing language techniques were used to extract features from it that were used to train, validate and test three models built with different techniques. The best performance was produced by the Bidirectional Long Short Term Memory (LSTM) Neural Network model.

  • How to Detect AI-Generated Texts? [paper]

    Trung Nguyen, Amartya Hatua and Andrew H. Sung.

    Present a method to distinguish synthetically generated text (SGT) from human-written text (HWT). The method includes methods for dataset creation, feature engineering, dataset comparison, and result analysis.

  • A Simple yet Efficient Ensemble Approach for AI-generated Text Detection. [paper]

    Harika Abburi, Kalyani Roy, Michael Suesserman, Nirmala Pudota, Balaji Veeramani, Edward Bowen and Sanmitra Bhattacharya.

    Propose a simple yet efficient solution to this problem by ensembling predictions from multiple constituent LLMs. Study the influence that the training data from individual LLMs have on model performance. Find that substituting commercially-restrictive GPT data with data generated from other open language models is a feasible alternative when developing generative text detectors.

  • Detecting and Unmasking AI-Generated Texts through Explainable Artificial Intelligence using Stylistic Features. [paper]

    Aditya Shah, Prateek Ranka, Urmi Dedhia, Shruti Prasad, Siddhi Muni and Kiran Bhowmick.

    The research delves into parameters such as syllable count, word length, sentence structure, functional word usage, and punctuation ratios to detect AI-generated text. Furthermore, the research integrates Explainable AI (xAI) techniques—LIME and SHAP—to enhance the interpretability of machine learning model predictions.

  • DEMASQ: Unmasking the ChatGPT Wordsmith. [paper]

    Kavita Kumari, Alessandro Pegoraro, Hossein Fereidooni and Ahmad-Reza Sadeghi.

    Propose an effective ChatGPT detector named DEMASQ, which accurately identifies ChatGPT-generated content. The method addresses two critical factors: (i) the distinct biases in text composition observed in human- and machinegenerated content and (ii) the alterations made by humans to evade previous detection methods.

  • AuthentiGPT: Detecting Machine-Generated Text via Black-Box Language Models Denoising. [paper]

    Zhen Guo and Shangdi Yu.

    Present AuthentiGPT, an efficient classifier that distinguishes between machine-generated and human-written texts. Under the assumption that humanwritten text resides outside the distribution of machine-generated text, AuthentiGPT leverages a black-box LLM to denoise input text with artificially added noise, and then semantically compares the denoised text with the original to determine if the content is machine-generated.

  • How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect LLM-Generated Text Detection. [paper]

    Ryuto Koike, Masahiro Kaneko and Naoaki Okazaki.

    Discover that even a task-oriented constraint in instruction can cause the inconsistent performance of current detectors to the generated texts. Show that the detection performance variance of the current detector on texts generated by instruction with each task-oriented constraint is up to 20 times larger than the variance caused by generating texts multiple times and paraphrasing the instruction.

  • Word-Level Interpretation of Chatgpt Detector Based on Classification Contribution. [paper]

    Dekun Chen.

    Present a method to show the contribution of words to model predictions. Specifically, this study considers information from classification weight vectors, semantic vectors, and embedded input word vectors for a more complete interpretation of detector LLM.

  • A Ship of Theseus: Curious Cases of Paraphrasing in LLM-Generated Texts. [paper]

    Nafis Irtiza Tripto, Saranya Venkatraman, Dominik Macko, Robert Moro, Ivan Srba, Adaku Uchendu, Thai Le and Dongwon Lee.

    Explore the diverse notions of authorship regarding paraphrasing in author attribution and AI text detection, incorporating philosophical concepts like the Ship of Theseus paradox. Suggest that authorship should be task-dependent, and we substantiate our results with relevant theoretical and philosophical perspectives.

  • Spot the Bot: Distinguishing Human-Written and Bot-Generated Texts Using Clustering and Information Theory Techniques. [paper]

    Vasilii Gromov and Quynh Nhu Dang.

    Propose a bot identification algorithm that is based on unsupervised learning techniques and does not depend on a large amount of labelled data. By combining findings in semantic analysis by clustering (crisp and fuzzy) and information techniques, we construct a robust model that detects a generated text for different types of bot.

  • Beyond Turing: A Comparative Analysis of Approaches for Detecting Machine-Generated Text. [paper]

    Muhammad Farid Adilazuarda, Nikolaos Nektarios Arkoulis and Oleksii Chumakov.

    Present the IMGTB framework, which simplifies the benchmarking of machine-generated text detection methods by easy integration of custom (new) methods and evaluation datasets.

  • IMGTB: A Framework for Machine-Generated Text Detection Benchmarking. [paper]

    Michal Spiegel and Dominik Macko.

    Offer an in-depth evaluation of three distinct methods: traditional shallow learning, Language Model (LM) fine-tuning, and Multilingual Model fine-tuning. These approaches are rigorously tested on a wide range of machine-generated texts, providing a benchmark of their competence in distinguishing between human-authored and machine-authored linguistic constructs.

  • Machine-Generated Text Detection using Deep Learning. [paper]

    Raghav Gaggar, Ashish Bhagchandani and Harsh Oza.

    Create a wellconstructed labeled dataset encompassing texts originated by both humans and ChatGPT (GPT3.5 turbo) from five distinct sources. Additionally, design and train multiple classifiers to differentiate between texts generated by humans and ChatGPT.

  • Token Prediction as Implicit Classification to Identify LLM-Generated Text. [paper]

    Yutian Chen, Hao Kang, Vivian Zhai, Liangze Li, Rita Singh and Bhiksha Raj.

    Instead of adding an additional classification layer to a base LM, we reframe the classification task as a next-token prediction task and directly fine-tune the base LM to perform it. Utilize the Text-to-Text Transfer Transformer (T5) model as the backbone for the experiments.

  • The Effectiveness of Free Software for Detecting AI-Generated Writing. [paper]

    Gregory Price and Marc Sakellarios.

    Explore the effectiveness of freely available AI detection software in identifying AI-generated content and its implications for teacher evaluation of student work. Compare our manual analysis with results from five different free AI writing detectors.

  • Beat LLMs at Their Own Game: Zero-Shot LLM-Generated Text Detection via Querying ChatGPT. [paper]

    Biru Zhu, Lifan Yuan, Ganqu Cui, Yangyi Chen, Chong Fu, Bingxiang He, Yangdong Deng, Zhiyuan Liu, Maosong Sun and Ming Gu.

    Design a zero-shot black-box method for detecting LLM-generated texts. The key idea is to revise the text to be detected using the ChatGPT model. Our method is based on the intuition that the ChatGPT model will make fewer revisions to LLM-generated texts than it does to human-written texts, because the texts generated by LLMs are more in accord with the generation logic and statistical patterns learned by LLMs like ChatGPT.

  • Evaluating the Efficacy of Hybrid Deep Learning Models in Distinguishing AI-Generated Text. [paper]

    Finbarrs Oketunji.

    Investigate the use of cutting-edge hybrid deep learning models to accurately differentiate between AI-generated text and human writing. Combining sophisticated neural networks, the custom model enabled it to detect nuanced differences between AI and human content.

  • STADEE: STAtistics-based DEEp Detection of Machine Generated Text. [paper]

    Zheng Chen and Huming Liu.

    Introduce STADEE, a STAtistics-based DEEp detection method that integrates essential statistical features of text with a sequence-based deep classifier. Utilize various statistical features, such as the probability, rank, cumulative probability of each token, as well as the information entropy of the distribution at each position.

  • Classification of Human- and AI-Generated Texts for English, French, German, and Spanish. [paper]

    Kristina Schaaff, Tim Schlippe and Lorenz Mindner.

    analyze features to classify human- and AI-generated text for English, French, German and Spanish and compare them across languages. Investigate two scenarios: (1) The detection of text generated by AI from scratch, and (2) the detection of text rephrased by AI.

  • ALTA Shared Task 2023

    • Overview of the 2023 ALTA Shared Task: Discriminate between Human-Written and Machine-Generated Text. [paper]

      Diego Molla, Haolan Zhan, Xuanli He and Qiongkai Xu.

    • A Prompt in the Right Direction: Prompt Based Classification of Machine-Generated Text Detection. [paper]

      Rinaldo Gagiano and Lin Tian.

    • Automatic Detection of Machine-Generated Text Using Pre-Trained Language Models. [paper]

      Yunhao Fang.

    • An Ensemble Based Approach To Detecting LLM-Generated Texts. [paper]

      Ahmed EI-Sayed and Omar Nasr.

    • Feature-Level Ensemble Learning for Robust Synthetic Text Detection with DeBERTaV3 and XLM-RoBERTa. [paper]

      Saman Sarker Joy and Tanusree Das Aishi.

    • Stacking the Odds: Transformer-Based Ensemble for AI-Generated Text Detection. [paper]

      Duke Nguyen, Khaing Myat Noe Naing and Aditya Joshi.

  • HuRef: HUman-REadable Fingerprint for Large Language Models. [paper]

    Boyi Zeng, Chenghu Zhou, Xinbing Wang and Zhouhan Lin.

    Introduce HuRef, a human-readable fingerprint for LLMs that uniquely identifies the base model without exposing model parameters or interfering with training. The method generates a dog image as an identity fingerprint for an LLM, where the dog’s appearance strongly indicates the LLM’s base model. Specifically, if the LLM is adapted from another base model, the generated dog highly resembles that model; otherwise if trained independently from scratch, it exhibits a unique dog image distinct from other models.

  • Deep dive into language traits of AI-generated Abstracts. [paper]

    Vikas Kumar, Amisha Bharti, Devanshu Verma and Vasudha Bhatnagar.

    Attempt to detect the Abstracts generated by ChatGPT, which are much shorter in length and bounded. Extract the text’s semantic and lexical properties and observe that traditional machine learning models can confidently detect these Abstracts.

  • Assaying on the Robustness of Zero-Shot Machine-Generated Text Detectors. [paper]

    Yi-Fan Zhang, Zhang Zhang, Liang Wang, Tieniu Tan and Rong Jin.

    Existing zero-shot detectors, typically designed for specific tasks or topics, often assume uniform testing scenarios, limiting their practicality. Uncover a significant correlation between topics and detection performance. Delve into the influence of topic shifts on zero-shot detectors.

  • Assessing AI Detectors in Identifying AI-Generated Code: Implications for Education. [paper]

    Wei Hung Pan, Ming Jie Chok, Jonathan Leong Shan Wong, Yung Xin Shin, Yeong Shian Poon, Zhou Yang, Chun Yong Chong, David Lo and Mei Kuan Lim.

    Collect a dataset comprising 5,069 samples, with each sample consisting of a textual description of a coding problem and its corresponding human-written Python solution codes. Results demonstrate that existing AIGC Detectors perform poorly in distinguishing between human-written code and AI-generated code.

  • Between Lines of Code: Unraveling the Distinct Patterns of Machine and Human Programmers. [paper]

    Yuling Shi, Hongyu Zhang, Chengcheng Wan and Xiaodong Gu.

    Propose a novel machine-generated code detection method called DetectCodeGPT, which improves DetectGPT by capturing the distinct structural patterns of code. Diverging from conventional techniques that depend on external LLMs for perturbations, DetectCodeGPT perturbs the code corpus by strategically inserting spaces and newlines, ensuring both efficacy and efficiency.

  • Enhancing Robustness of LLM-Synthetic Text Detectors for Academic Writing: A Comprehensive Analysis. [paper]

    Zhicheng Dou, Yuchen Guo, Ching-Chun Chang, Huy H. Nguyen and Isao Echizen.

    Present a comprehensive analysis of the impact of prompts on the text generated by LLMs and highlight the potential lack of robustness in one of the current state-of-the-art GPT detectors. Propose a reference-based Siamese detector named Synthetic-Siamese which takes a pair of texts, one as the inquiry and the other as the reference.

  • Deciphering Textual Authenticity: A Generalized Strategy through the Lens of Large Language Semantics for Detecting Human vs. Machine-Generated Text. [paper]

    Mazal Bethany, Brandon Wherry, Emet Bethany, Nishant Vishwamitra and Peyman Najafirad.

    Undertake a systematic study on the detection of machine-generated text in real-world scenarios. Study the effectiveness of state-of-the-art approaches and find that they are severely limited against text produced by diverse generators and domains in the real world. Introduce a novel system, T5LLMCipher, for detecting machine-generated text using a pretrained T5 encoder combined with LLM embedding sub-clustering to address the text produced by diverse generators and domains in the real world.

  • ALMs: Authorial Language Models for Authorship Attribution. [paper]

    Weihang Huang, Akira Murakami and Jack Grieve.

    Introduce an authorship attribution method called Authorial Language Models (ALMs) that involves identifying the most likely author of a questioned document based on the perplexity of the questioned document calculated for a set of causal language models fine-tuned on the writings of a set of candidate author.

  • Fine-tuning Large Language Models for Multigenerator, Multidomain, and Multilingual Machine-Generated Text Detection. [paper]

    Feng Xiong, Thanet Markchom, Ziwei Zheng, Subin Jung, Varun Ojha and Huizhi Liang.

    Two methods: 1) using traditional machine learning (ML) with natural language preprocessing (NLP) for feature extraction, and 2) fine-tuning LLMs for text classification. The results show that transformer models, particularly LoRA-RoBERTa, exceed traditional ML methods in effectiveness, with majority voting being particularly effective in multilingual contexts for identifying machine-generated texts.

  • Beyond Lexical Boundaries: LLM-Generated Text Detection for Romanian Digital Libraries. [paper]

    Melania Nitu and Mihai Dascalu.

    Introduce a corpus of approximately 60 k Romanian documents, including human-written samples as well as generated texts using six distinct Large Language Models (LLMs) and three different generation methods. Introduce two techniques for discerning machine-generated texts. The study also includes a text similarity analysis between human-authored and artificially generated texts, coupled with a SHAP analysis.

  • Does DetectGPT Fully Utilize Perturbation? Selective Perturbation on Model-Based Contrastive Learning Detector would be Better. [paper]

    Shengchao Liu, Xiaoming Liu, Yichen Wang, Zehua Cheng, Chengzhengxu Li, Zhaohan Zhang, Yu Lan and Chao Shen.

    DetectGPT's random perturbation strategy might introduce noise, and its logit regression module harms the generalizability and applicability of individual or small-batch inputs. Propose a novel detector, PECOLA, which uses selective strategy perturbation to relieve the important information loss caused by random masking, and multi-pair contrastive learning to capture the implicit pattern information during perturbation, facilitating few-shot performance.

  • LLM-Detector: Improving AI-Generated Chinese Text Detection with Open-Source LLM Instruction Tuning. [paper]

    Rongsheng Wang, Haoming Chen, Ruizhe Zhou, Han Ma, Yaofei Duan, Yanlan Kang, Songhua Yang, Baoyu Fan and Tao Tan.

    Create a dataset that mixed human-written sentences and sentences polished by LLMs. Propose LLM-Detector, a novel method for both document-level and sentence-level text detection through Instruction Tuning of LLMs. Our method leverages the wealth of knowledge LLMs acquire during pre-training, enabling them to detect the text they generate. Instruction tuning aligns the model's responses with the user's expected text detection tasks.

  • Detecting Generated Native Ads in Conversational Search. [paper]

    Sebastian Schmidt, Ines Zelch, Janek Bevendorff, Benno Stein, Matthias Hagen and Martin Potthast.

    This paper investigates whether LLMs can be used as a countermeasure against generated native ads, i.e., to block them. Compile a large dataset of ad-prone queries and of generated answers with automatically integrated ads to experiment with fine-tuned sentence transformers and state-of-the-art LLMs on the task of recognizing the ads.

  • Ten Words Only Still Help: Improving Black-Box AI-Generated Text Detection via Proxy-Guided Efficient Re-Sampling. [paper]

    Yuhui Shi, Qiang Sheng, Juan Cao, Hao Mi, Beizhe Hu and Danding Wang.

    Propose to estimate word generation probabilities as pseudo white-box features via multiple re-sampling to help improve AIGT detection under the black-box setting. Design POGER, a proxy-guided efficient re-sampling method, which selects a small subset of representative words for performing multiple re-sampling in black-box AIGT detection.

  • Threads of Subtlety: Detecting Machine-Generated Texts Through Discourse Motifs. [paper]

    Zae Myung Kim, Kwang Hee Lee, Preston Zhu, Vipul Raheja and Dongyeop Kang.

    This paper delves into the inquiry of identifying discernible and unique linguistic properties in texts that were written by humans, particularly uncovering the underlying discourse structures of texts beyond their surface structures. Introducing a novel methodology, we leverage hierarchical parse trees and recursive hypergraphs to unveil distinctive discourse patterns in texts produced by both LLMs and humans.

  • Spotting AI’s Touch: Identifying LLM-Paraphrased Spans in Text. [paper]

    Anonymous authors.

    Propose a novel detection framework, paraphrased text span detection (PTD), aiming to identify paraphrased text spans within a text. Different from text-level detection, PTD takes in the full text and assigns each of the sentences with a score indicating the paraphrasing degree. Construct a dedicated dataset, PASTED, for paraphrased text span detection.

  • Who is the Writer?Identifying the Generative Model by Writing Style. [paper]

    Anonymous authors.

    Present the Identify the Writer by Writing Style (IWWS) model, a novel approach designed to identify the writing styles of human and generative model. Propose a comprehensive dataset, HumanGenTextify. The IWWS model not only achieves high accuracy in text source identification but also provides insights into the distinctive writing styles that characterize human and model-generated texts.

  • SemEval-2024 Task 8

    • Team QUST at SemEval-2024 Task 8: A Comprehensive Study of Monolingual and Multilingual Approaches for Detecting AI-generated Text. [paper]

      Xiaoman Xu, Xiangrun Li, Taihang Wang, Jianxiang Tian and Ye Jiang.

    • HU at SemEval-2024 Task 8A: Can Contrastive Learning Learn Embeddings to Detect Machine-Generated Text? [paper]

      Shubhashis Roy Dipta and Sadat Shahriar.

    • KInIT at SemEval-2024 Task 8: Fine-tuned LLMs for Multilingual Machine-Generated Text Detection. [paper]

      Michal Spiegel and Dominik Macko.

    • RFBES at SemEval-2024 Task 8: Investigating Syntactic and Semantic Features for Distinguishing AI-Generated and Human-Written Texts. [paper]

      Mohammad Heydari Rad, Farhan Farsi, Shayan Bali, Romina Etezadi and Mehrnoush Shamsfard.

    • SemEval-2024 Task 8: Weighted Layer Averaging RoBERTa for Black-Box Machine-Generated Text Detection. [paper]

      Ayan Datta, Aryan Chandramania and Radhika Mamidi.

    • MasonTigers at SemEval-2024 Task 8: Performance Analysis of Transformer-based Models on Machine-Generated Text Detection. [paper]

      Sadiya Sayara Chowdhury Puspo, Md Nishat Raihan, Dhiman Goswami, Al Nahian Bin Emran, Amrita Ganguly and Ozlem Uzuner.

    • TrustAI at SemEval-2024 Task 8: A Comprehensive Analysis of Multi-domain Machine Generated Text Detection Techniques. [paper]

      Ashok Urlana, Aditya Saibewar, Bala Mallikarjunarao Garlapati, Charaka Vinayak Kumar, Ajeet Kumar Singh and Srinivasa Rao Chalamala.

    • AIpom at SemEval-2024 Task 8: Detecting AI-produced Outputs in M4. [paper]

      Alexander Shirnin, Nikita Andreev, Vladislav Mikhailov and Ekaterina Artemova.

    • TM-TREK at SemEval-2024 Task 8: Towards LLM-Based Automatic Boundary Detection for Human-Machine Mixed Text. [paper]

      Xiaoyan Qu and Xiangfeng Meng.

    • AISPACE at SemEval-2024 task 8: A Class-balanced Soft-voting System for Detecting Multi-generator Machine-generated Text. [paper]

      Renhua Gu and Xiangfeng Meng.

    • PetKaz at SemEval-2024 Task 8: Can Linguistics Capture the Specifics of LLM-generated Text?. [paper]

      Kseniia Petukhova, Roman Kazakov and Ekaterina Kochmar.

  • Machine-generated Text Localization. [paper]

    Zhongping Zhang, Wenda Qin and Bryan A. Plummer.

    Provide the first in-depth study of MGT that localizes the portions of a document that were machine generated. A key challenge in the MGT localization task is that short spans of text, e.g., a single sentence, provides little information indicating if it is machine generated due to its short length. Leverage contextual information, where we predict whether multiple sentences are machine or human written at once.

  • Whose LLM is it Anyway? Linguistic Comparison and LLM Attribution for GPT-3.5, GPT-4 and Bard. [paper]

    Ariel Rosenfeld and Teddy Lazebnik.

    Through a comprehensive linguistic analysis, compare the vocabulary, Part-Of-Speech (POS) distribution, dependency distribution, and sentiment of texts generated by three of the most popular LLMS today (GPT-3.5, GPT-4, and Bard) to diverse inputs. The results point to significant linguistic variations which, in turn, enable us to attribute a given text to its LLM origin with a favorable 88% accuracy using a simple off-the-shelf classification model.

  • Technical Report on the Checkfor.ai AI-Generated Text Classifier. [paper]

    Bradley Emi and Max Spero.

    Present the Checkfor.ai text classifier, a transformer-based neural network trained to distinguish text written by large language models from text written by humans. Propose a training algorithm, hard negative mining with synthetic mirrors, that enables the classifier to achieve orders of magnitude lower false positive rates on high-data domains such as reviews.

  • Machine-Generated Text Detection and Attribution. [paper]

    Petr Šigut.

    Analyse various ways to detect AI-generated content, particularly when encountering content in languages other than English. Test how three AI detection tools, Compilatio, Turnitin and GPT-2 Output Detector, perform with Czech, Slovak and English texts generated by ChatGPT.

  • Deep Learning Detection Method for Large Language Models-Generated Scientific Content. [paper]

    Bushra Alhijawi, Rawan Jarrar, Aseel AbuAlRub and Arwa Bader.

    Present a novel ChatGPT-generated scientific text detection method, AI-Catcher. AI-Catcher integrates two deep learning models, multilayer perceptron (MLP) and convolutional neural networks (CNN). In addition, a new ChatGPT-Generated scientific text dataset is collected to enhance AI-generated text detection tools, AIGTxt.

  • Towards Detecting AI-Generated Text within Human-AI Collaborative Hybrid Texts. [paper]

    Zijie Zeng, Shiqi Liu, Lele Sha, Zhuang Li, Kaixun Yang, Sannyuya Liu, Dragan Gašević and Guanliang Chen.

    Explore the challenge of sentence-level AI-generated text detection within human-AI collaborative hybrid texts. Utilize the CoAuthor dataset, which includes diverse, realistic hybrid texts generated through the collaboration between human writers and an intelligent writing system in multi-turn interactions. Adopt a two-step, segmentation-based pipeline: (i) detect segments within a given hybrid text where each segment contains sentences of consistent authorship, and (ii) classify the authorship of each identified segment.

  • Adaptive Ensembles of Fine-Tuned Transformers for LLM-Generated Text Detection. [paper]

    Zhixin Lai, Xuesheng Zhang and Suiyao Chen.

    Reveal that single transformer-based classifiers achieved decent performance on in-distribution dataset but limited generalization ability on out-of-distribution dataset. To improve it, combine the individual classifiers models using adaptive ensemble algorithms. The results indicate the effectiveness, good generalization ability, and great potential of adaptive ensemble algorithms in LLM-generated text detection.

  • EAGLE: A Domain Generalization Framework for AI-generated Text Detection. [paper]

    Amrita Bhattacharjee, Raha Moraffah, Joshua Garland and Huan Liu.

    Propose a domain generalization framework for the detection of AI-generated text from unseen target generators. The proposed framework, EAGLE, leverages the labeled data that is available so far from older language models and learns features invariant across these generators, in order to detect text generated by an unknown target generator. EAGLE learns such domain-invariant features by combining the representational power of self-supervised contrastive learning with domain adversarial training.

  • Detecting AI-Generated Code Assignments Using Perplexity of Large Language Models. [paper]

    Zhenyu Xu and Victor S. Sheng.

    Address the challenge of determining whether a student's code assignment was generated by a language model. First, the proposed method identifies AI-generated code by leveraging targeted masking perturbation paired with comperhesive scoring. Second, utilize a fine-tuned CodeBERT to fill in the masked portions, producing subtle modified samples. Then, integrate the overall perplexity, variation of code line perplexity, and burstiness into a unified score. In this scoring scheme, a higher rank for the original code suggests it's more likely to be AI-generated.

  • MUGC: Machine Generated versus User Generated Content Detection. [paper]

    Yaqi Xie, Anjali Rawal, Yujing Cen, Dixuan Zhao, Sunil K Narang and Shanu Sushmita.

    Undertake a comparative evaluation of eight traditional machine-learning algorithms to distinguish between machine-generated and human-generated data across three diverse datasets: Poems, Abstracts, and Essays.

  • The Impact of Prompts on Zero-Shot Detection of AI-Generated Text. [paper]

    Kaito Taguchi, Yujie Gu and Kouichi Sakurai.

    Introduce an evaluative framework to empirically analyze the impact of prompts on the detection accuracy of AI-generated text. Assess various zero-shot detectors using both white-box detection, which leverages the prompt, and black-box detection, which operates without prompt information. Experiments reveal the significant influence of prompts on detection accuracy.

  • Detection of AI-Generated Text Using Large Language Model. [paper]

    Manish Prajapati, Santos Kumar Baliarsingh, Chinmayee Dora, Ashutosh Bhoi, Jhalak Hota and Jasaswi Prasad Mohanty.

    Leverage the machine learning (ML) models to reliably identify whether an essay is authored by a human being or by an LLM. Using diverse texts and unknown generative models, we replicate typical scenarios to encourage feature learning across models.

Detector Attack

  • Attacking Neural Text Detectors. [paper]

    Max Wolff, Stuart Wolff.

    Present two classes of black-box attacks on neural text detectors, one which randomly replaces characters with homoglyphs, and the other a simple scheme to purposefully misspell words. The homoglyph and misspelling attacks decrease a popular neural text detector's recall on neural text from 97.44% to 0.26% and 22.68%, respectively.

  • Adversarial Robustness of Neural-Statistical Features in Detection of Generative Transformers. [paper]

    Evan Crothers, Nathalie Japkowicz, Herna Viktor and Paula Branco.

    Assess each model’s relative performance at classifying computer-generated text. Evaluat models for robustness in the presence of text adversarial attacks. Find that statistical features are considerably more robust to adversarial attack than Transformer-derived features.

  • Exploring Semantic Perturbations on Grover. [paper]

    Pranav Kulkarni, Ziqing Ji, Yan Xu, Marko Neskovic and Kevin Nolan.

    Find that Grover-generated substitution, GPT2-generated substitution, length changing, and perturbations informed with the model’s embedding table were all effective to some degree in fooling Grover’s detection.

  • Can AI-Generated Text be Reliably Detected? [paper]

    Vinu Sankar Sadasivan, Aounon Kumar, Sriram Balasubramanian, Wenxiao Wang and Soheil Feizi.

    Show that paraphrasing attacks can break detectors using the watermarking schemes as well as neural network-based detectors and zero-shot classifiers. Even LLMs protected by watermarking schemes can be vulnerable against spoofing attacks.

  • Paraphrasing evades detectors of AI-generated text, but retrieval is an effective defense. [paper]

    Kalpesh Krishna, Yixiao Song, Marzena Karpinska, John Wieting and Mohit Iyyer.

    Present DIPPER, a discourse paraphrase generation model. DIPPER paraphrases easily evade these detectors while approximately preserving input semantics. Propose a retrieval-based mechanism to defend against such paraphrase attacks. Experiments show that this retrieval defense significantly outperforms baseline detectors on paraphrased text, and is effective at scale.

  • GPT detectors are biased against non-native English writers. [paper]

    Weixin Liang, Mert Yuksekgonul, Yining Mao, Eric Wu and James Zou.

    Find that several widely-used GPT detectors consistently misclassify non-native English writing samples as AI-generated, whereas native writing samples are accurately identified. Demonstrate that simple prompting strategies can not only mitigate this bias but also effectively bypass GPT detectors.

  • Stochastic Parrots Looking for Stochastic Parrots: LLMs are Easy to Fine-Tune and Hard to Detect with other LLMs. [paper]

    Da Silva Gameiro Henrique, Andrei Kucharavy and Rachid Guerraoui.

    Show that generative LLM detection with a discriminator LLM is impossible if the attacker has access to the reference "human" dataset used to train the discriminator LLM. Simply fine-tuning the dataset and using prompts from it leads to a complete failure of the discriminator to learn the difference between machine and human-generated texts, even in a setting where all LLM outputs are correctly labeled.

  • Large Language Models can be Guided to Evade AI-Generated Text Detection. [paper]

    Ning Lu, Shengcai Liu, Rui He, Qi Wang and Ke Tang.

    Reveal that with the aid of carefully crafted prompts, LLMs can effectively evade detection systems. Propose a novel Substitution-based In-Context example Optimization method (SICO) to automatically generate such prompts. On three real-world tasks where LLMs can be misused, SICO successfully enables ChatGPT to evade six existing detectors, causing a significant 0.54 AUC drop on average.

  • Red Teaming Language Model Detectors with Language Models. [paper]

    Zhouxing Shi, Yihan Wang, Fan Yin, Xiangning Chen, Kai-Wei Chang and Cho-Jui Hsieh.

    Systematically test the reliability of the existing detectors, by designing two types of attack strategies to fool the detectors: 1) replacing words with their synonyms based on the context; 2) altering the writing style of generated text. Reveal that these attacks effectively compromise the performance of all tested detectors.

  • Evade ChatGPT Detectors via A Single Space. [paper]

    Shuyang Cai and Wanyun Cui.

    Find that detectors do not effectively discriminate the semantic and stylistic gaps between human-generated and AI-generated conten. Instead, the “subtle differences”, such as an extra space, become crucial for detection. Based on this discovery, propose the SpaceInfi strategy to evade detection.

  • Language Model Detectors Are Easily Optimized Against. [paper]

    Charlotte Nicks, Eric Mitchell, Rafael Rafailov, Archit Sharma, Christopher Manning, Chelsea Finn and Stefano Ermon.

    demonstrate a data-efficient attack that fine-tunes language models to confuse existing detectors, leveraging recent developments in reinforcement learning of language models. Similar to traditional adversarial attacks, we find that this increase in ‘detector evasion’ generalizes to other detectors not used during training.

  • Black-Box Adversarial Attacks Against Language Model Detector. [paper]

    Huyen Ha, Duc Tran and Dukyun Kim.

    Conduct a systematic analysis on the resilience of the LM Detector against eight black-box adversarial attack methods. Propose a new technique, called StrictPWWS that introduces the semantic similarity constraint into the conventional Probability Weighted Word Saliency (PWWS).

  • Hidding the Ghostwriters: An Adversarial Evaluation of AI-Generated Student Essay Detection. [paper]

    Xinlin Peng, Ying Zhou, Ben He, Le Sun and Yingfei Sun.

    Construct AIG-ASAP, an AI-generated student essay dataset, employing a range of text perturbation methods that are expected to generate high-quality essays while evading detection. Through empirical experiments, assess the performance of current AIGC detectors on the AIG-ASAP dataset. The results reveal that the existing detectors can be easily circumvented using straightforward automatic adversarial attacks.

  • Authorship Obfuscation in Multilingual Machine-Generated Text Detection. [paper]

    Dominik Macko, Robert Moro, Adaku Uchendu, Ivan Srba, Jason Samuel Lucas, Michiharu Yamashita, Nafis Irtiza Tripto, Dongwon Lee, Jakub Simko and Maria Bielikova.

    Machine-generated text (MGT) detection is susceptible to authorship obfuscation (AO) methods, such as paraphrasing. Benchmark the performance of 10 well-known AO methods, attacking 37 MGT detection methods against MGTs in 11 languages. The results indicate that all tested AO methods can cause detection evasion in all tested languages, where homoglyph attacks are especially successful.

  • ALISON: Fast and Effective Stylometric Authorship Obfuscation. [paper]

    Eric Xing, Saranya Venkatraman, Thai Le and Dongwon Lee.

    Propose a practical Authorship Obfuscation method, ALISON. Demonstrate that ALISON can effectively prevent four SOTA AA methods from accurately determining the authorship of ChatGPT-generated texts, all while minimally changing the original text semantics.

  • Hidding the Ghostwriters: An Adversarial Evaluation of AI-Generated Student Essay Detection. [paper]

    Xinlin Peng, Ying Zhou, Ben He, Le Sun and Yingfei Sun.

    Construct AIG-ASAP, an AI-generated student essay dataset, employing a range of text perturbation methods that are expected to generate high-quality essays while evading detection. Assess the performance of current AIGC detectors on the AIG-ASAP dataset. The results reveal that the existing detectors can be easily circumvented using straightforward automatic adversarial attacks.

  • Stumbling Blocks: Stress Testing the Robustness of Machine-Generated Text Detectors Under Attacks. [paper]

    Yichen Wang, Shangbin Feng, Abe Bohan Hou, Xiao Pu, Chao Shen, Xiaoming Liu, Yulia Tsvetkov and Tianxing He.

    The goal of our study is to stress test the detectors' robustness to malicious attacks under realistic scenarios. We comprehensively study the robustness of popular machine-generated text detectors under attacks from diverse categories: editing, paraphrasing, prompting, and co-generating. Our attacks assume limited access to the generator LLMs, and we compare the performance of detectors on different attacks under different budget levels. Our experiments reveal that almost none of the existing detectors remain robust under all the attacks, and all detectors exhibit different loopholes. Averaging all detectors, the performance drops by 35% across all attacks.

  • Token-Ensemble Text Generation: On Attacking the Automatic AI-Generated Text Detection. [paper]

    Fan Huang, Haewoon Kwak and Jisun An.

    Propose a novel token-ensemble generation strategy to challenge the robustness of current AI-content detection approaches. Explore the ensemble attack strategy by completing the prompt with the next token generated from random candidate LLMs. Find the token-ensemble approach significantly drops the performance of AI-content detection models.

  • Navigating the Shadows: Unveiling Effective Disturbances for Modern AI Content Detectors. [paper]

    Anonymous authors.

    Simulate real-world scenarios in both informal and professional writing, exploring the out-of-the-box performance of current detectors. Construct 12 black-box text perturbation methods to assess the robustness of current detection models across various perturbation granularities. Through adversarial learning experiments, investigate the impact of perturbation data augmentation on the robustness of AI-text detectors.

  • Can Watermarks Survive Translation? On the Cross-lingual Consistency of Text Watermark for Large Language Models. [paper]

    Zhiwei He, Binglin Zhou, Hongkun Hao, Aiwei Liu, Xing Wang, Zhaopeng Tu, Zhuosheng Zhang and Rui Wang.

    Introduce the concept of ''cross-lingual consistency'' in text watermarking, which assesses the ability of text watermarks to maintain their effectiveness after being translated into other languages. Propose a Cross-lingual Watermark Removal Attack (CWRA) to bypass watermarking by first obtaining a response from an LLM in a pivot language, which is then translated into the target language.

  • Attacking LLM Watermarks by Exploiting Their Strengths. [paper]

    Qi Pang, Shengyuan Hu, Wenting Zheng and Virginia Smith.

    Show that desirable properties shared by existing LLM watermarking systems such as quality preservation, robustness, and public detection APIs can in turn make these systems vulnerable to various attacks. Rigorously study potential attacks in terms of common watermark design choices, and propose best practices and defenses for mitigation -- establishing a set of practical guidelines for embedding and detection of LLM watermarks.

  • Humanizing Machine-Generated Content: Evading AI-Text Detection through Adversarial Attack. [paper]

    Ying Zhou, Ben He and Le Sun.

    Propose a framework for a broader class of adversarial attacks, designed to perform minor perturbations in machine-generated content to evade detection. Consider two attack settings: white-box and black-box, and employ adversarial learning in dynamic scenarios to assess the potential enhancement of the current detection model's robustness against such attacks.

Benchmark

  • TURINGBENCH: A Benchmark Environment for Turing Test in the Age of Neural Text Generation. [paper]

    Adaku Uchendu, Zeyu Ma, Thai Le, Rui Zhang and Dongwon Lee.

    TuringBench is a benchmark environment that contains: (1)Benchmark tasks- Turing Test (i.e., human vs. machine) and Authorship Attribution: (i.e., who is the author of this texts?) (2) Datasets (Binary and Multi-class settings) (3) Website with leaderboard.

  • A Benchmark Corpus for the Detection of Automatically Generated Text in Academic Publications. [paper]

    Vijini Liyanage, Davide Buscaldi and Adeline Nazarenko.

    Propose two datasets comprised of artificially generated research content: a completely synthetic dataset and a partial text substitution dataset.

  • MGTBench: Benchmarking Machine-Generated Text Detection. [paper]

    Xinlei He, Xinyue Shen, Zeyuan Chen, Michael Backes and Yang Zhang.

    Consider 6 metric-based detection methods and 2 model-based detection methods under ChatGPT. And integrate the detection methods as well as datasets into a modular-designed framework named MGTBench.

  • M4: Multi-generator, Multi-domain, and Multi-lingual Black-Box Machine-Generated Text Detection. [paper]

    Yuxia Wang, Jonibek Mansurov, Petar Ivanov, Jinyan Su, Artem Shelmanov, Akim Tsvigun, Chenxi Whitehouse, Osama Mohammed Afzal, Tarek Mahmoud, Alham Fikri Aji and Preslav Nakov.

    Introduce a large-scale benchmark M4, which is multigenerator, multi-domain, and multi-lingual corpus for machine-generated text detection. Using the dataset, experiment with a number of methods and show that it is challenging for detectors to generalize well on unseen examples if they are either from different domains or are generated by different large language models.

  • Distinguishing Fact from Fiction: A Benchmark Dataset for Identifying Machine-Generated Scientific Papers in the LLM Era. [paper]

    Edoardo Mosca, Mohamed Hesham Ibrahim Abdalla, Paolo Basso, Margherita Musumeci and Georg Groh.

    Introduce a novel benchmark dataset, containing human-written and machine-generated scientific papers from SCIgen, GPT-2, GPT-3, ChatGPT, and Galactica. Experiment with four distinct classifiers as a baseline for detecting the authorship of scientific text.

  • Med-MMHL: A Multi-Modal Dataset for Detecting Human- and LLM-Generated Misinformation in the Medical Domain. [paper]

    Yanshen Sun, Jianfeng He, Shuo Lei, Limeng Cui and Chang-Tien Lu.

    Present Med-MMHL, a novel multi-modal misinformation detection dataset in a general medical domain encompassing multiple diseases. Med-MMHL not only incorporates human-generated misinformation but also includes misinformation generated by LLMs like ChatGPT.

  • An Empirical Study of AI Generated Text Detection Tools. [paper]

    Arslan Akram.

    Create a multidomain dataset for testing the state-of-the-art APIs and tools for detecting artificially generated information used by universities and other research institutions. Six different artificial intelligence (AI) text identification systems, including "GPTkit," "GPTZero," "Originality," "Sapling," "Writer," and "Zylalab," have accuracy rates between 55.29 and 97.0%.

  • Who Said That? Benchmarking Social Media AI Detection. [paper]

    Wanyun Cui, Linqiu Zhang, Qianle Wang and Shuyang Cai.

    Introduce SAID (Social media AI Detection), a novel benchmark developed to assess AI-text detection models' capabilities in real social media platforms. It incorporates real AI-generate text from popular social media platforms like Zhihu and Quora. Present a new user-oriented AI-text detection challenge focusing on the practicality and effectiveness of identifying AI-generated text based on user information and multiple responses.

  • MULTITuDE: Large-Scale Multilingual Machine-Generated Text Detection Benchmark. [paper]

    Dominik Macko, Robert Moro, Adaku Uchendu, Jason Samuel Lucas, Michiharu Yamashita, Matúš Pikuliak, Ivan Srba, Thai Le, Dongwon Lee, Jakub Simko and Maria Bielikova.

    Introduce MULTITuDE, a novel benchmarking dataset for multilingual machine-generated text detection comprising of 74,081 authentic and machine-generated texts in 11 languages (ar, ca, cs, de, en, es, nl, pt, ru, uk, and zh) generated by 8 multilingual LLMs.

  • HANSEN: Human and AI Spoken Text Benchmark for Authorship Analysis. [paper]

    Nafis Tripto, Adaku Uchendu, Thai Le, Mattia Setzu, Fosca Giannotti and Dongwon Lee.

    Introduce the largest benchmark for spoken texts - HANSEN (Human ANd ai Spoken tExt beNchmark). HANSEN encompasses meticulous curation of existing speech datasets accompanied by transcripts, alongside the creation of novel AI-generated spoken text datasets. Together, it comprises 17 human datasets, and AI-generated spoken texts created using 3 prominent LLMs: ChatGPT, PaLM2, and Vicuna13B.

  • LLM-as-a-Coauthor: The Challenges of Detecting LLM-Human Mixcase. [paper]

    Chujie Gao, Dongping Chen, Qihui Zhang, Yue Huang, Yao Wan and Lichao Sun.

    Collect mixcase instances generated from multiple daily text-editing scenarios and composed MixSet, the first dataset dedicated to studying these mixed modification scenarios. Conduct experiments to evaluate the efficacy of popular MGT detectors. Reveal that existing detectors struggle to identify mixcase as a separate class or MGT.

  • AiGen-FoodReview: A Multimodal Dataset of Machine-Generated Restaurant Reviews and Images on Social Media. [paper]

    Alessandro Gambetti and Qiwei Han.

    Leveraging OpenAI's GPT-4-Turbo and DALL-E-2 models, craft AiGen-FoodReview, a multi-modal dataset of 20,144 restaurant review-image pairs divided into authentic and machine-generated. Explore unimodal and multimodal detection models, achieving 99.80% multimodal accuracy with FLAVA.

  • M4GT-Bench: Evaluation Benchmark for Black-Box Machine-Generated Text Detection. [paper]

    Yuxia Wang, Jonibek Mansurov, Petar Ivanov, Jinyan Su, Artem Shelmanov, Akim Tsvigun, Osama Mohanned Afzal, Tarek Mahmoud, Giovanni Puccetti, Thomas Arnold, Alham Fikri Aji, Nizar Habash, Iryna Gurevych and Preslav Nakov.

    Address this problem by introducing a new benchmark involving multilingual, multi-domain and multi-generator for MGT detection -- M4GT-Bench. It is collected for three task formulations: (1) mono-lingual and multi-lingual binary MGT detection; (2) multi-way detection identifies which particular model generates the text; and (3) human-machine mixed text detection, where a word boundary delimiting MGT from human-written content should be determined.

  • GPT-generated Text Detection: Benchmark Dataset and Tensor-based Detection Method. [paper]

    Zubair Qazi, William Shiao and Evangelos E. Papalexakis.

    Present GPT Reddit Dataset (GRiD), a novel Generative Pretrained Transformer (GPT)-generated text detection dataset designed to assess the performance of detection models in identifying generated responses from ChatGPT. The dataset consists of a diverse collection of context-prompt pairs based on Reddit, with human-generated and ChatGPT-generated responses. Propose GpTen, a novel tensor-based GPT text detection method.

Watermarking

  • Adversarial Watermarking Transformer: Towards Tracing Text Provenance with Data Hiding. [paper]

    Sahar Abdelnabi and Mario Fritz.

  • Tracing Text Provenance via Context-Aware Lexical Substitution. [paper]

    Xi Yang, Jie Zhang, Kejiang Chen, Weiming Zhang, Zehua Ma, Feng Wang and Nenghai Yu.

  • On Information Hiding in Natural Language Systems. [paper]

    Geetanjali Bihani, Julia Taylor Rayz.

  • Protecting Intellectual Property of Language Generation APIs with Lexical Watermark. [paper]

    Xuanli He, Qiongkai Xu, Lingjuan Lyu, Fangzhao Wu and Chenguang Wang.

  • DeepHider: A Covert NLP Watermarking Framework Based on Multi-task Learning. [paper]

    Long Dai, Jiarong Mao, Xuefeng Fan and Xiaoyi Zhou.

  • CATER: Intellectual Property Protection on Text Generation APIs via Conditional Watermarks. [paper]

    Xuanli He, Qiongkai Xu, Yi Zeng, Lingjuan Lyu, Fangzhao Wu, Jiwei Li and Ruoxi Jia.

  • A Watermark for Large Language Models. [paper]

    John Kirchenbauer, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers and Tom Goldstein.

  • Protecting Language Generation Models via Invisible Watermarking. [paper]

    Xuandong Zhao, Yu-Xiang Wang and Lei Li.

  • Robust Natural Language Watermarking through Invariant Features. [paper]

    KiYoon Yoo, Wonhyuk Ahn, Jiho Jang and Nojun Kwak.

  • DeepTextMark: A Deep Learning-Driven Text Watermarking Approach for Identifying Large Language Model Generated Text. [paper]

    Travis Munyer, Abdullah Tanvir, Arjon Das and Xin Zhong.

  • Watermarking Text Generated by Black-Box Language Models. [paper]

    Xi Yang, Kejiang Chen, Weiming Zhang, Chang Liu, Yuang Qi, Jie Zhang, Han Fang and Nenghai Yu.

  • Are You Copying My Model? Protecting the Copyright of Large Language Models for EaaS via Backdoor Watermark. [paper]

    Wenjun Peng, Jingwei Yi, Fangzhao Wu, Shangxi Wu, Bin Zhu, Lingjuan Lyu, Binxing Jiao, Tong Xu, Guangzhong Sun and Xing Xie.

  • Watermarking Text Data on Large Language Models for Dataset Copyright Protection. [paper]

    Yixin Liu, Hongsheng Hu, Xuyun Zhang and Lichao Sun.

  • Who Wrote this Code? Watermarking for Code Generation. [paper]

    Taehyun Lee, Seokhee Hong, Jaewoo Ahn, Ilgee Hong, Hwaran Lee, Sangdoo Yun, Jamin Shin and Gunhee Kim.

  • Undetectable Watermarks for Language Models. [paper]

    Miranda Christ, Sam Gunn and Or Zamir.

  • Baselines for Identifying Watermarked Large Language Models. [paper]

    Leonard Tang, Gavin Uberti and Tom Shlomi.

  • On the Reliability of Watermarks for Large Language Models. [paper]

    John Kirchenbauer, Jonas Geiping, Yuxin Wen, Manli Shu, Khalid Saifullah, Kezhi Kong, Kasun Fernando, Aniruddha Saha, Micah Goldblum and Tom Goldstein.

  • Provable Robust Watermarking for AI-Generated Text. [paper]

    Xuandong Zhao, Prabhanjan Vijendra Ananth, Lei Li and Yu-Xiang Wang.

  • Watermarking Conditional Text Generation for AI Detection: Unveiling Challenges and a Semantic-Aware Watermark Remedy. [paper]

    Yu Fu, Deyi Xiong and Yue Dong.

  • Robust Distortion-free Watermarks for Language Models. [paper]

    Rohith Kuditipudi, John Thickstun, Tatsunori Hashimoto and Percy Liang.

  • Three Bricks to Consolidate Watermarks for Large Language Models. [paper]

    Pierre Fernandez, Antoine Chaffin, Karim Tit, Vivien Chappelier and Teddy Furon.

  • Towards Codable Text Watermarking for Large Language Models. [paper]

    Lean Wang, Wenkai Yang, Deli Chen, Hao Zhou, Yankai Lin, Fandong Meng, Jie Zhou and Xu Sun.

  • An Unforgeable Publicly Verifiable Watermark for Large Language Models. [paper]

    Aiwei Liu, Leyi Pan, Xuming Hu, Shu'ang Li, Lijie Wen, Irwin King and Philip S. Yu.

  • Advancing Beyond Identification: Multi-bit Watermark for Large Language Models. [paper]

    KiYoon Yoo, Wonhyuk Ahn and Nojun Kwak.

  • Unbiased Watermark for Large Language Models. [paper]

    Zhengmian Hu, Lichang Chen, Xidong Wu, Yihan Wu, Hongyang Zhang and Heng Huang.

  • Necessary and Sufficient Watermark for Large Language Models. [paper]

    Yuki Takezawa, Ryoma Sato, Han Bao, Kenta Niwa and Makoto Yamada.

  • SemStamp: A Semantic Watermark with Paraphrastic Robustness for Text Generation. [paper]

    Abe Bohan Hou, Jingyu Zhang, Tianxing He, Yichen Wang, Yung-Sung Chuang, Hongwei Wang, Lingfeng Shen, Benjamin Van Durme, Daniel Khashabi and Yulia Tsvetkov.

  • A Semantic Invariant Robust Watermark for Large Language Models. [paper]

    Aiwei Liu, Leyi Pan, Xuming Hu, Shiao Meng and Lijie Wen.

  • DiPmark: A Stealthy, Efficient and Resilient Watermark for Large Language Models. [paper]

    Yihan Wu, Zhengmian Hu, Hongyang Zhang and Heng Huang.

  • WASA: WAtermark-based Source Attribution for Large Language Model-Generated Data. [paper]

    Jingtan Wang, Xinyang Lu, Zitong Zhao, Zhongxiang Dai, Chuan-Sheng Foo, See-Kiong Ng and Bryan Kian Hsiang Low.

  • On the Learnability of Watermarks for Language Models. [paper]

    Chenchen Gu, Xiang Lisa Li, Percy Liang and Tatsunori Hashimoto.

  • Double-I Watermark: Protecting Model Copyright for LLM Fine-tuning. [paper]

    Shen Li, Liuyi Yao, Jinyang Gao, Lan Zhang and Yaliang Li.

  • I Know You Did Not Write That! A Sampling Based Watermarking Method for Identifying Machine Generated Text. [paper]

    Kaan Efe Keleş, Ömer Kaan Gürbüz and Mucahid Kutlu.

  • Watermarking LLMs with Weight Quantization. [paper]

    Linyang Li, Botian Jiang, Pengyu Wang, Ke Ren, Hang Yan and Xipeng Qiu.

  • Embarrassingly Simple Text Watermarks. [paper]

    Ryoma Sato, Yuki Takezawa, Han Bao, Kenta Niwa and Makoto Yamada.

  • REMARK-LLM: A Robust and Efficient Watermarking Framework for Generative Large Language Models. [paper]

    Ruisi Zhang, Shehzeen Samarah Hussain, Paarth Neekhara and Farinaz Koushanfar.

  • Publicly Detectable Watermarking for Language Models. [paper]

    Jaiden Fairoze, Sanjam Garg, Somesh Jha, Saeed Mahloujifar, Mohammad Mahmoody and Mingyuan Wang.

  • Watermarks in the Sand: Impossibility of Strong Watermarking for Generative Models. [paper]

    Hanlin Zhang, Benjamin L. Edelman, Danilo Francati, Daniele Venturi, Giuseppe Ateniese and Boaz Barak.

  • WaterBench: Towards Holistic Evaluation of Watermarks for Large Language Models. [paper]

    Shangqing Tu, Yuliang Sun, Yushi Bai, Jifan Yu, Lei Hou and Juanzi Li.

  • A Robust Semantics-based Watermark for Large Language Model against Paraphrasing. [paper]

    Jie Ren, Han Xu, Yiding Liu, Yingqian Cui, Shuaiqiang Wang, Dawei Yin and Jiliang Tang.

  • WatME: Towards Lossless Watermarking Through Lexical Redundancy. [paper]

    Liang Chen, Yatao Bian, Yang Deng, Deng Cai, Shuaiyi Li, Peilin Zhao and Kam-fai Wong.

  • Improving the Generation Quality of Watermarked Large Language Models via Word Importance Scoring. [paper]

    Yuhang Li, Yihan Wang, Zhouxing Shi and Cho-Jui Hsieh.

  • Performance Trade-offs of Watermarking Large Language Models. [paper]

    Anirudh Ajith, Sameer Singh and Danish Pruthi.

  • Mark My Words: Analyzing and Evaluating Language Model Watermarks. [paper]

    Julien Piet, Chawin Sitawarin, Vivian Fang, Norman Mu and David Wagner.

  • Towards Optimal Statistical Watermarking. [paper]

    Baihe Huang, Banghua Zhu, Hanlin Zhu, Jason D. Lee, Jiantao Jiao and Michael I. Jordan.

  • Cross-Attention Watermarking of Large Language Models. [paper]

    Folco Bertini Baldassini, Huy H. Nguyen, Ching-Chung Chang and Isao Echizen.

  • Adaptive Text Watermark for Large Language Models. [paper]

    Yepeng Liu and Yuheng Bu.

  • Instructional Fingerprinting of Large Language Models. [paper]

    Jiashu Xu, Fei Wang, Mingyu Derek Ma, Pang Wei Koh, Chaowei Xiao and Muhao Chen.

  • Provably Robust Multi-bit Watermarking for AI-generated Text via Error Correction Code. [paper]

    Wenjie Qu, Dong Yin, Zixin He, Wei Zou, Tianyang Tao, Jinyuan Jia and Jiaheng Zhang.

  • Permute-and-Flip: An optimally robust and watermarkable decoder for LLMs. [paper]

    Xuandong Zhao, Lei Li and Yu-Xiang Wang.

  • Resilient Watermarking for LLM-Generated Codes. [paper]

    Boquan Li, Mengdi Zhang, Peixin Zhang, Jun Sun and Xingmei Wang.

  • Proving membership in LLM pretraining data via data watermarks. [paper]

    Johnny Tian-Zheng Wei, Ryan Yixiang Wang and Robin Jia.

  • k-SemStamp: A Clustering-Based Semantic Watermark for Detection of Machine-Generated Text. [paper]

    Abe Bohan Hou, Jingyu Zhang, Yichen Wang, Daniel Khashabi and Tianxing He.

  • Duwak: Dual Watermarks in Large Language Models. [paper]

    Chaoyi Zhu, Jeroen Galjaard, Pin-Yu Chen and Lydia Y. Chen.

  • Bypassing LLM Watermarks with Color-Aware Substitutions. [paper]

    Qilong Wu and Varun Chandrasekaran.

  • GumbelSoft: Diversified Language Model Watermarking via the GumbelMax-trick. [paper]

    Jiayi Fu, Xuandong Zhao, Ruihan Yang, Yuansen Zhang, Jiangjie Chen and Yanghua Xiao.

  • Watermarking Makes Language Models Radioactive. [paper]

    Tom Sander, Pierre Fernandez, Alain Durmus, Matthijs Douze and Teddy Furon.

  • Multi-Bit Distortion-Free Watermarking for Large Language Models. [paper]

    Massieh Kordi Boroujeny, Ya Jiang, Kai Zeng and Brian Mark.

  • EmMark: Robust Watermarks for IP Protection of Embedded Quantized Large Language Models. [paper]

    Ruisi Zhang and Farinaz Koushanfar.

  • Token-Specific Watermarking with Enhanced Detectability and Semantic Coherence for Large Language Models. [paper]

    Mingjia Huo, Sai Ashish Somayajula, Youwei Liang, Ruisi Zhang, Farinaz Koushanfar and Pengtao Xie.

  • WaterMax: breaking the LLM watermark detectability-robustness-quality trade-off. [paper]

    Eva Giboulot and Furon Teddy.

  • Lost in Overlap: Exploring Watermark Collision in LLMs. [paper]

    Yiyang Luo, Ke Lin and Chao Gu.

  • Learning to Watermark LLM-generated Text via Reinforcement Learning. [paper]

    Xiaojun Xu, Yuanshun Yao and Yang Liu.

  • Towards Better Statistical Understanding of Watermarking LLMs. [paper]

    Zhongze Cai, Shang Liu, Hanzhao Wang, Huaiyang Zhong and Xiaocheng Li.

  • An Entropy-based Text Watermarking Detection Method. [paper]

    Yijian Lu, Aiwei Liu, Dianzhi Yu, Jingjing Li and Irwin King.

  • WaterJudge: Quality-Detection Trade-off when Watermarking Large Language Models. [paper]

    Piotr Molenda, Adian Liusie and Mark J. F. Gales.

  • A Statistical Framework of Watermarks for Large Language Models: Pivot, Detection Efficiency and Optimal Rules. [paper]

    Xiang Li, Feng Ruan, Huiyuan Wang, Qi Long and Weijie J. Su.

  • Topic-based Watermarks for LLM-Generated Text. [paper]

    Alexander Nemecek, Yuzhou Jiang and Erman Ayday.

Other Related Work

  • MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers. [paper]

    Krishna Pillutla, Swabha Swayamdipta, Rowan Zellers, John Thickstun, Sean Welleck, Yejin Choi and Zaid Harchaoui.

    Develop MAUVE, a comparison measure for open-ended text generation, which directly compares the learnt distribution from a text generation model to the distribution of human-written text using divergence frontiers.

  • Is GPT-3 Text Indistinguishable from Human Text? Scarecrow: A Framework for Scrutinizing Machine Text. [paper]

    Yao Dou, Maxwell Forbes, Rik Koncel-Kedziorski, Noah A. Smith and Yejin Choi.

    Develop Scarecrow, a methodology for eliciting categorical judgements of errors in machine-generated text from crowd workers. Quantify measurable gaps between human authored text and generations from models of several sizes.

  • Automatic Detection of Entity-Manipulated Text using Factual Knowledge. [paper]

    Ganesh Jawahar, Muhammad Abdul-Mageed and Laks Lakshmanan.

    Present an approach that can detect entity-manipulated text articles. The proposed detector exploits explicit factual knowledge from a knowledge base to overcome the limitations of relying only on stylometric signals. Constructed challenging manipulated datasets by considering various entity replacement strategies, including with random selection and GPT-2 generation.

  • Unraveling the Mystery of Artifacts in Machine Generated Text. [paper]

    Jiashu Pu, Ziyi Huang, Yadong Xi, Guandan Chen, Weijie Chen and Rongsheng Zhang.

    Reveal characteristics and possible manifestations of artifacts. Locate artifacts mainly based on the performance variation of the classifier against different corruptions.

  • On the Possibilities of AI-Generated Text Detection. [paper]

    Souradip Chakraborty, Amrit Singh Bedi, Sicheng Zhu, Bang An, Dinesh Manocha and Furong Huang.

    Provide evidence that it should almost always be possible to detect the AI-generated text unless the distributions of human and machine generated texts are exactly the same over the entire support.

  • Towards an Understanding and Explanation for Mixed-Initiative Artificial Scientific Text Detection. [paper]

    Luoxuan Weng, Minfeng Zhu, Kam Kwai Wong, Shi Liu, Jiashun Sun, Hang Zhu, Dongming Han and Wei Chen.

    A formative study that identifies critical distinctions and summarizes design requirements for artificial scientific text detection. Propose a novel workflow and a prototype system that combines human intelligence with machine learning techniques to facilitate artificial text detection.

  • AI, write an essay for me: A large-scale comparison of human-written versus ChatGPT-generated essays. [paper]

    Steffen Herbold, Annette Hautli-Janisz, Ute Heuer, Zlata Kikteva and Alexander Trautsch.

    Compare human-written versus ChatGPT-generated argumentative student essays. The results demonstrate that ChatGPT generates essays that are rated higher for quality than human-written essays. The writing style of the AI models exhibits linguistic characteristics that are different from those of the human-written essays.

  • Bot or Human? Detecting ChatGPT Imposters with A Single Question. [paper]

    Hong Wang, Xuan Luo, Weizhi Wang and Xifeng Yan.

    Propose a framework named FLAIR, Finding Large Language Model Authenticity via a Single Inquiry and Response, to detect conversational bots in an online manner. The questions that can effectively differentiate human users from bots are divided into two categories: those that are easy for humans but difficult for bots, and those that are easy for bots but difficult for humans.

  • On the Risk of Misinformation Pollution with Large Language Models. [paper]

    Yikang Pan, Liangming Pan, Wenhu Chen, Preslav Nakov, Min-Yen Kan and William Yang Wang.

    To mitigate the harm caused by LLM-generated misinformation, explore three defense strategies: prompting, misinformation detection, and majority voting. Initial results show promising trends for these defensive strategies.

  • Social bot detection in the age of ChatGPT: Challenges and opportunities. [paper]

    Emilio Ferrara.

    Suggest potentially promising opportunities and research directions in social bot detection, including the use of generative agents for synthetic data generation, testing and evaluation.

  • Are you in a Masquerade? Exploring the Behavior and Impact of Large Language Model Driven Social Bots in Online Social Networks. [paper]

    Siyu Li, Jin Yang and Kui Zhao.

    Provide a comprehensive analysis of LLMs-driven social bots. Collect and publicly release the first activity behavior dataset of LLMs-driven social bots, named Masquerade-23.

  • "I Slept Like a Baby": Using Human Traits To Characterize Deceptive ChatGPT and Human Text. [paper]

    Salvatore Giorgi, David M. Markowitz, Nikita Soni, Vasudha Varadarajan, Siddharth Mangalik and H. Andrew Schwartz.

    Characterize differences between (a) truthful text written by humans, (b) intentionally deceptive text written by humans, and (c) inherently deceptive text written by state-of-the-art language models (ChatGPT). Building on these differences, train a classifier using only the thirteen human traits to automatically discriminate between truthful and deceptive language, with a classification AUC of up to 0.966.

  • Anatomy of an AI-powered malicious social botnet. [paper]

    Kai-Cheng Yang and Filippo Menczer.

    Unveil the emergence of LLM-powered social bots. Combine the 1,140 bot accounts with the 1,140 human ones results in a benchmark dataset: the fox8-23 dataset. Apply state-of-the-art LLM-content detectors and find they cannot effectively distinguish between human and LLM-powered bots in the wild.

  • Detecting The Corruption Of Online Questionnaires By Artificial Intelligence. [paper]

    Benjamin Lebrun, Sharon Temtsin, Andrew Vonasch and Christoph Bartneck.

    This study tested if text generated by an AI for the purpose of an online study can be detected by both humans and automatic AI detection systems. While humans were able to correctly identify authorship of text above chance level (76 percent accuracy), their performance was still below what would be required to ensure satisfactory data quality. Researchers currently have to rely on the disinterest of bad actors to successfully use open-ended responses as a useful tool for ensuring data quality.

  • Spot the bot: Coarse-Grained Partition of Semantic Paths for Bots and Humans. [paper]

    Vasilii A. Gromov and Alexandra S. Kogan.

    This paper focuses on comparing structures of the coarse-grained partitions of semantic paths for human-written and bot-generated texts. Compare the clusterizations of datasets of n-grams from literary texts and texts generated by several bots. The hypothesis is that the structures and clusterizations are different. The research supports the hypothesis.

  • Contra generative AI detection in higher education assessments. [paper]

    Cesare G. Ardito.

    This paper presents a critical analysis of generative Artificial Intelligence (AI) detection tools in higher education assessments. Explore the effectiveness, vulnerabilities, and ethical implications of AI detection tools in the context of preserving academic integrity. This paper advocates for a strategic shift towards robust assessment methods and educational policies that embrace generative AI usage while ensuring academic integrity and authenticity in assessments.

  • Perfect detection of computer-generated text faces fundamental challenges. [paper]

    Martin Májovský, Martin Černý, David Netuka and Tomáš Mikolov.

    Propose a focus on ethical guidelines rather than outright prohibitions, emphasizing that technological solutions should complement, not replace, the core ethical principles of scientific publishing.

  • TextMachina: Seamless Generation of Machine-Generated Text Datasets. [paper]

    Areg Mikael Sarvazyan, José Ángel González and Marc Franco-Salvador.

    Introduce TextMachina, a modular and extensible Python framework, designed to aid in the creation of high-quality, unbiased datasets to build robust models for MGT-related tasks such as detection, attribution, or boundary detection.

  • To Burst or Not to Burst: Generating and Quantifying Improbable Text. [paper]

    Kuleen Sasse, Samuel Barham, Efsun Sarioglu Kayi and Edward W. Staley.

    While large language models (LLMs) are extremely capable at text generation, their outputs are still distinguishable from human-authored text. Introduce a new metric, recoverability, to highlight differences between human and machine text; and propose a new sampling technique, burst sampling, designed to close this gap.

  • Lying Blindly: Bypassing ChatGPT's Safeguards to Generate Hard-to-Detect Disinformation Claims at Scale. [paper]

    Freddy Heppell, Mehmet E. Bakir and Kalina Bontcheva.

    This study explores the capability of ChatGPT to generate unconditioned claims about the war in Ukraine, an event beyond its knowledge cutoff, and evaluates whether such claims can be differentiated by human readers and automated tools from human-written ones.

  • Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews. [paper]

    Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuandong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang, Daniel A. McFarland and James Y. Zou.

    Present an approach for estimating the fraction of text in a large corpus which is likely to be substantially modified or produced by a large language model (LLM). Apply this approach to a case study of scientific peer review in AI conferences that took place after the release of ChatGPT: ICLR 2024, NeurIPS 2023, CoRL 2023 and EMNLP 2023. Results suggest that between 6.5% and 16.9% of text submitted as peer reviews to these conferences could have been substantially modified by LLMs.

  • Quantitative Analysis of AI-Generated Texts in Academic Research: A Study of AI Presence in Arxiv Submissions using AI Detection Tool. [paper]

    Arslan Akram.

    This study analyze a method that can see purposely manufactured content that academic organizations use to post on Arxiv. A dataset was created using physics, mathematics, and computer science articles. The statistical analysis shows that Originality.ai is very accurate, with a rate of 98%.

  • GenAI Detection Tools, Adversarial Techniques and Implications for Inclusivity in Higher Education. [paper]

    Mike Perkins, Jasper Roe, Binh H. Vu, Darius Postma, Don Hickerson, James McGaughran and Huy Q. Khuat.

    Investigate the efficacy of six major Generative AI (GenAI) text detectors when confronted with machine-generated content that has been modified using techniques designed to evade detection by these tools (n=805). The results demonstrate that the detectors' already low accuracy rates (39.5%) show major reductions in accuracy (17.4%) when faced with manipulated content, with some techniques proving more effective than others in evading detection.

  • Contrasting Linguistic Patterns in Human and LLM-Generated News Text. [paper]

    Alberto Muñoz-Ortiz, Carlos Gómez-Rodríguez and David Vilares.

    Conduct a quantitative analysis contrasting human-written English news text with comparable large language model (LLM) output from from six different LLMs that cover three different families and four sizes in total. The analysis spans several measurable linguistic dimensions, including morphological, syntactic, psychometric, and sociolinguistic aspects. The results reveal various measurable differences between human and AI-generated texts.

  • Mapping the Increasing Use of LLMs in Scientific Papers. [paper]

    Weixin Liang, Yaohui Zhang, Zhengxuan Wu, Haley Lepp, Wenlong Ji, Xuandong Zhao, Hancheng Cao, Sheng Liu, Siyu He, Zhi Huang, Diyi Yang, Christopher Potts, Christopher D Manning and James Y. Zou.

    Conduct the first systematic, large-scale analysis across 950,965 papers published between January 2020 and February 2024 on the arXiv, bioRxiv, and Nature portfolio journals, using a population-level statistical framework to measure the prevalence of LLM-modified content over time.

  • David versus Goliath: Can Machine Learning Detect LLM-Generated Text? A Case Study in the Detection of Phishing Emails. [paper]

    Francesco Greco, Giuseppe Desolda, Andrea Esposito and Alessandro Carelli.

    Detecting whether an email is generated by LLMs is critical. This work aims to address this problem using traditional, lightweight machine learning models that are easy to interpret and require fewer computational resources. This approach allows users to understand why an email is AI-generated, improving their decision-making in the case of phishing emails.

Demos & Products

Datasets

Dateset Reference Release Time Description Generation Model Link
gpt-2-output-dataset [paper] 2019 250K random samples and 250K samples generated with Top-K 40 truncation GPT-2 [repo]
Grover generation [paper] 2019 15,000 human-written and 10,000 generated news article Grover [repo]
fakenews machine data [paper] 2019 2000 newsQA answers generated by Grover and 2000 news articles edited by GPT-2 Grover, GPT-2 [home] [file]
Deepfake bot submissions [paper] 2019 795 human-written comments and 1,001 generated comments GPT-2 [home]
NeuralNews [paper] 2020 128k articles, 27% of which are generated text Grover [home] [file]
TweepFake [paper] 2020 25,572 tweets half human and half bots generated varied [home]
Mixed NLG [paper] 2020 8 different Transformer-based LM, 1066 texts from each of the models and 1066 human-written texts varied [repo]
Generated Headlines [paper] 2020 72,401 real and 414,373 generated headlines in "defender" set, 179, 880 real and 517, 932 generated in “attacker" set GPT-2 [repo]
TuringBench [paper] 2021 168,612 articles from 19 machine text-generators and 1 human varied [repo]
GeneratedTextDetection [paper] 2022 100 artificially generated research papers aligned to the original abstract and 100 combinations of original content and machine generated sentences GPT-2 [repo]
COVID Generated Data [paper] 2022 50k examples from GPT-3, about 60k examples each from 4 sizes of GPT-2 and GPT-NEO and 3 decoding technique GPT-2, GPT-NEO, GPT-3 (davinci) [repo] [file]
In-The-Wild [paper] 2022 3,887 human-written articles and 3,887 synthetic articles from AI-Writer, ArticleForge, Kafkai and RedditBot varied [repo]
RoFT Dataset [paper] 2022 21,646 annotations for human to find the boundary between human-written text and machine-generated text GPT-3 (davinci) [repo]
Self-Instruct [paper] 2022 52k instructions, paired with 82K instance inputs and outputs GPT-3 (davinci) [repo]
ChatGPT Generated Text Detection Corpus [paper] 2023 126 humans essays and 126 nonhumans essays ChatGPT [repo]
Human ChatGPT Comparison Corpus (HC3) [paper] 2023 58,546 human answers and 26,903 ChatGPT answrers in English, 22,259 human answers and 17,522 ChatGPT answrers in Chinese ChatGPT [repo]
MGTBench [paper] 2023 Human and ChatGPT answers of 817 questions from TruthfulQA, 1,000 questions from SQuAD1 and 1,000 questions from NarrativeQA ChatGPT [repo]
Alpaca 2023 52K instruction-following data generated by text-davinci-003 GPT-3.5 (text-davinci-003) [home] [repo]
Cleaned Alpaca 2023 A cleaned version of the original Alpaca Dataset GPT-3.5 (text-davinci-003) [repo] [repo]
GPT4All [paper] 2023 437,605 prompt-generation pairs GPT-3.5 (gpt-3.5-turbo) [repo] [repo]
Instruction Tuning with GPT-4 [paper] 2023 52K instruction-following data generated by GPT-4 GPT-4 [home] [repo]
ArguGPT [paper] 2023 4,115 human-written essays and 4,038 machine-generated essays produced by 7 GPT models varied [repo]
DeepfakeTextDetect [paper] 2023 154,078 human-written texts and 294,381 machine-generated texts produced by 27 LLMs varied [repo]
M4 [paper] 2023 122k human–machine parallel data and over 10M non-parallel human-written texts varied [repo]
Ghostbuster [paper] 2023 All datasets were generated using gpt-3.5-turbo, with the exception of the paragraph-level data, which was generated using text-davinci-003’s insert feature GPT-3.5 (gpt-3.5-turbo, text-davinci-003) [repo]
SnifferBench [paper] 2023 6k human written texts, and every 6k texts from GPT2, GPT-Neo, GPT-J, and 12k texts from GPT3 models, which is 36k texts in total varied [repo]
HPPT [paper] 2023 6050 pairs of abstracts and corresponding polished versions from ACL anthology ChatGPT [repo]
IDMGSP [paper] 2023 16k human-written texts and 13k machine-generated scientific papers varied [repo]
Med-MMHL [paper] 2023 41,365 human-written sentences and 17,608 LLM-generated scentences from the medical domain GPT-3.5 (gpt-3.5-turbo) [repo]
OpenGPTText [paper] 2023 A high quality dataset with approximately 30,000 text sample rephrased by gpt-3.5-turbo. GPT-3.5 (gpt-3.5-turbo) [repo] [file]
HC-Var [paper] 2023 Contains 4 different types of language tasks, each task covers 1 to 4 different topics. GPT-3.5 (gpt-3.5-turbo) [repo]
SAID [paper] 2023 87,215 human-written responses and 131,546 AI-generated responses from Zhihu and Quora Unknown [repo]
SeqXGPT-Bench [paper] 2023 30,000 documents generated by GPT-2, GPT-J, GPT-Neo, LLaMA, GPT-3.5-turbo and human. varied [repo]
MULTITuDE [paper] 2023 7992 human-written news texts in 11 languages, accompanied by 66089 texts generated by 8 large language models. varied [repo]
HANSEN [paper] 2023 Encompass existing speech datasets accompanied by transcripts, alongside the creation of novel AI-generated spoken text datasets. GPT-3.5 (gpt-3.5-turbo), PaLM 2 (chat-bison@001), Vicuna 13B [repo]
OpenLLMText [paper] 2023 Approximately 340,000 text samples from five sources. GPT-3.5, PaLM, LLaMA7B, GPT2-1B (GPT2 extra large) [repo]
CHEAT [paper] 2023 35,304 synthetic abstracts, with Generation, Polish, and Fusion as prominent representatives. GPT-3.5 (gpt-3.5-turbo) [repo]
MixSet [paper] 2024 A total of 3.6k mixcase instances. GPT-4, LLaMA-2 70B [repo]
GRiD [paper] 2024 1368 samples generated by GPT-3.5-turbo, whereas 5145 samples authored by human contributors. GPT-3.5 (gpt-3.5-turbo) [repo]

Shared Tasks