Skip to content
View jon-chun's full-sized avatar

Block or report jon-chun

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
jon-chun/README.md

Interdisciplinary AI Researcher

Jon A Chun
Co-Founder with Katherine Elkins @ github.com/KatherineElkins, Kenyon Human-Centered AI Colab


Contents


Overview

I am an interdisciplinary ML/AI researcher and educator specializing in bridging traditional academic divisions, AI research, industry practices, and emerging social topics including government regulation, ethics, and entrepreneurship. My research focuses on ML/AI approaches to language, narrative, emotion, cognition, and persuasion/deception, employing data science, statistical machine learning, and deep learning including NLP, LLM, and LMM. My work extends to eXplainable AI (XAI), fairness-accuracy-transparency-explainability (FATE), AI ethical auditing, and AI regulation.

As an entrepreneur and innovator, I've contributed to diverse fields from network security and education to finance, insurance, and healthcare across environments ranging from Silicon Valley startups to higher education institutions. I've published pioneering interdisciplinary AI research on storytelling and emotion in major venues including Narrative and the Modern Language Association (MLA). In 2016, I co-founded the first human-centered AI curriculum integrating domain experts from literature, music, political science, and economics. We established the term "AI Digital Humanities" and have guided over 300 original ML/AI DH projects as of 2025, achieving approximately 100k downloads (as of January 2026) from leading institutions worldwide. Currently, I serve as a co-principal investigator for the US NIST AI Safety Institute representing the Modern Language Association and collaborate with the IBM-Notre Dame Tech Ethics Lab researching LLM prediction capabilities.

My prior experience includes co-founding the world's largest privacy and anonymity website, backed by In-Q-Tel. As CEO, I led a successful company pivot, co-authored patents on the first browser-based VPN appliance, and navigated the acquisition by Symantec, where I later directed the product's successful relaunch. Previously, I served as CIO for Silicon Valley's premier return-to-work firm and as CTO for international startups in Latin America and Japan. During medical school, I was an American Heart Association research fellow, publishing on gene therapy and pioneering web-based EMR systems in the American Medical Informatics Journal. My early career included roles as DELL's first US-based Japanese localization engineer, SEMATECH's initial Japanese patent analyst, and positions in Tokyo financial reporting, Lawrence Berkeley Labs' synchrotron facility (ALS), and Computer Associates' aerospace IT division.

Languages: English (native), Spanish (US Foreign Service Exam certified), Japanese (JLPT certified), French (professional working proficiency)

Professional interests: AI advancement, collaborative innovation, creative engineering, functional design, technical presentations, and enterprise solutions. Previous leadership experience includes coaching roles in baseball, soccer, wrestling, robotics, and Destination Imagination.


Research

My interdisciplinary research spans an unusually wide range of both STEM and non-STEM domains. Rather than maintaining a traditional focus on a single research specialty, I combine universal theories, sound engineering principles, diverse industry experience, and collaborations with domain experts to cover topics ranging from AI regulation to multimodal Affective AI to state-of-the-art AI model ethical benchmarking among others.

Early Career & Industry

  • Conducted research at Lawrence Berkeley Labs' Advanced Light Source Lab as an undergraduate researcher
  • Helped build out Asia-Pacific Financial Network from Tokyo office of Bloomberg provider McGraw-Hill
  • Served as a semiconductor research analyst for SEMATECH, leveraging Japanese language skills during graduate studies
  • Set up DELL's first Japanese testbed outside Japan (Austin, TX)
  • Built first resume submission/search website for the second largest recruiting firm Adecco
  • Co-authored genomics and medical informatics papers during medical school
  • Built first Workers Comp claims management web system for largest Silicon Valley return-to-work management firm
  • Fortune 500 Director of Development rebranding and launching one of the first VPN appliances within one year

Entrepreneurship & Innovation

  • Co-founded Brazilian-based international fashion, music, and culture lifestyle website (US, Japan, and Latin America)
  • CTO for logistics and computer supplier startup integrating supply chains across the US and Latin America
  • Co-founded and served as CEO of a successful Silicon Valley startup, B2C privacy/anonymity pivot to B2B enterprise security, acquired by world's largest security provider
  • Co-inventor on several VPN patents focusing on network security and anonymity
  • Judge for entrepreneur/startup competitions at UC Berkeley College of Engineering (graduate) and The Ohio State University (undergraduate)
  • Advisor for several student startups: EV in Sri Lanka, Crowdfunding in Venezuela, Culinary Social Network in US

Academic Leadership & AI Research

  • Pioneer in human-centered AI education, co-founding one of the world's first Interdisciplinary AI Colabs in the early 2010s
  • Published research in premier venues across AI, Computer Science, and Humanities:
    • Machine Learning: International Conference on Machine Learning (ICML)
    • Computational Linguistics: ACL's EMNLP/CoNLL
    • Computer Science: Frontiers in Computer Science
    • Humanities: Modern Language Association, Narrative, Cultural Analytics
    • Digital Humanities: Springer International Journal of Digital Humanities, International Journal of Humanities and Arts Computing

I created the open-source library SentimentArcs in 2019, at the time the largest ensemble for diachronic sentiment analysis and the basis for Katherine Elkins's The Shapes of Stories (Cambridge UP 2022). I presented some of the earliest GPT-2 story generation work at Narrative 2020 and have since published in Cultural Analytics and Narrative on AI and narrative. As of January 2025, I've mentored over 300 computational interdisciplinary ML/AI projects since 2017 across virtually every department of Kenyon College as part of the Integrated Program for Humane Studies and the Scientific Computing programs. I co-founded the AI Colab, the world's first human-centered AI curriculum at Kenyon, and our AI KDH research colab. I currently have research papers pending on using LLMs to compare multiple translations of Proust, multimodal (dialog+image) diachronic sentiment arcs in film, emotional hacking of LLM high-stakes decision-making, a novel benchmark on semantic similarity, IP infringement and creativity using Narrative theory, and an updated and expanding ethical audit of the leading LLMs. I'm also a co-author on an ICML position paper that was invited as an oral presentation this year in Vienna. My current research projects focus on AI persuasion, manipulation and deception as well as using LLMs for predictive analytics and decision-making on structured tabular data. (July 2024)

Academic Profiles:


Recent Highlights

Schmidt Sciences Humanities and AI Virtual Institute (HAVI)

"Archival Intelligence: Can AI Rescue Endangered Archives"

Schmidt Sciences Humanities and AI Virtual Institute (Co-PI on 1 of 23 Schmidt Sciences HAVI grants given to leading AI Research Teams worldwide)
Began: January 2026

Small archives hold one-of-a-kind historical materials that exist nowhere else. Those paying attention, however, are watching them disappear. Wealthy institutions can afford professional preservation, but small community archives cannot, and the impact is very real. Descendants are losing connection to their ancestors' stories, and communities are being severed from their own heritage. Even historians cannot tell complete stories, because primary sources are crumbling and in some cases, vanishing forever. A 1919 New Orleans newspaper (right) grows less readable each year, with the only remaining photo leaving a damaged, wrinkled corner in shadow. Small historical societies know what they're losing, but professional preservation equipment and archivists are too costly to be within reach.

A team of AI researchers, archival scientists, jazz historians, and New Orleans cultural experts is racing to stop this loss. The goal? Preservation using only a smartphone photo, with AI to recover portions lost to damage. The team is also building AI intelligence into recovered materials, connecting information across documents, languages, and formats so researchers and descendants can find what they're looking for and even facilitate new discoveries. This 18-month New Orleans pilot tackles the hardest cases by focusing on voices systematically excluded from historical archives: multilingual newspapers documenting Creole and Cajun communities as well as early jazz materials. If it works, thousands of small archives will finally have a way to save what's disappearing.


ArXiv Preprint (December 27, 2025)

Large language models (LLMs) are increasingly deployed in consequential decision-making settings, yet their robustness to benign prompt variation remains underexplored. In this work, we study whether LLMs maintain consistent ethical judgments across logically equivalent but syntactically different prompts, focusing on variations involving negation and conditional structure. We introduce Syntactic Framing Fragility (SFF), a robustness evaluation framework that isolates purely syntactic effects via Logical Polarity Normalization (LPN), enabling direct comparison of decisions across positive and negative framings without semantic drift. Auditing 23 state-of-the-art models spanning the U.S. and China as well as small U.S. open-source software models over 14 ethical scenarios and four controlled framings (39,975 decisions), we find widespread and statistically significant inconsistency: many models reverse ethical endorsements solely due to syntactic polarity, with open-source models exhibiting over twice the fragility of commercial counterparts. We further uncover extreme negation sensitivity, where some models endorse actions in 80-97% of cases when explicitly prompted with "should not." We show that eliciting chain-of-thought reasoning substantially reduces fragility, identifying a practical mitigation lever, and we map fragility across scenarios, finding higher risk in financial and business contexts than in medical scenarios. Our results demonstrate that syntactic consistency constitutes a distinct and critical dimension of ethical robustness, and we argue that SFF-style audits should be a standard component of safety evaluation for deployed LLMs. Code and results will be available on this URL.


Accepted: AAAI-26 LaMAS (Singapore, January 2026)
IBM-Notre Dame Tech Ethics Lab Grant

We introduce AgenticSimLaw, a role-structured, multi-agent debate framework that provides transparent and controllable test-time reasoning for high-stakes tabular decision-making tasks. Unlike black-box approaches, our courtroom-style orchestration explicitly defines agent roles (prosecutor, defense, judge), interaction protocols (7-turn structured debate), and private reasoning strategies, creating a fully auditable decision-making process. We benchmark this framework on young adult recidivism prediction using the NLSY97 dataset, comparing it against traditional chain-of-thought (CoT) prompting across almost 90 unique combinations of models and strategies. Our results demonstrate that structured multi-agent debate provides more stable and generalizable performance compared to single-agent reasoning, with stronger correlation between accuracy and F1-score metrics. Beyond performance improvements, AgenticSimLaw offers fine-grained control over reasoning steps, generates complete interaction transcripts for explainability, and enables systematic profiling of agent behaviors. While we instantiate this framework in the criminal justice domain to stress-test reasoning under ethical complexity, the approach generalizes to any deliberative, high-stakes decision task requiring transparency and human oversight. This work addresses key LLM-based multi-agent system challenges: organization through structured roles, observability through logged interactions, and responsibility through explicit non-deployment constraints for sensitive domains. Data, results, and code will be available on github.com under the MIT license.


"ESP-Ethical Audit: An Ethical Alignment Audit to Quantify Biased Decision-Making with Emotion and Syntactic Framing"

(In Review)

Autonomous agents and distributed AI networks are beginning to assist and even make decisions in real-world scenarios ranging from smart cities to HR. Ethical assessments are more urgent than ever, and our framework, ESP-Ethical Audit, offers a novel way to assess both strengths and vulnerabilities in current LLM systems. To date, there is surprisingly little quantitative assessment of underlying "ethical belief" systems in LLMs. This type of assessment samples the distribution of a binary classification task when LLMs are presented with complex and ethically-fraught scenarios. We interrogate whether decision-making by leading commercial and open-source LLM models align with human values and if so, whether that alignment reveals biases that typically occur in human decision-making. ESP-Ethical Audit measures confidence levels and explanatory frameworks across a variety of moral scenarios while testing perturbations along two dimensions: syntactic framing and empathic backstory. The former tests whether syntactic framing surfaces biases that mirror common framing biases documented in humans. The latter tests variations in backstory—both positive and negative—within two different relevancy scenarios. The first framework tests the introduction of facts that might be considered allowable in a US court of law; the other tests backstories that rely on emotional appeal. In all three framing cases, whether syntactic or backstory, we find more performant models more closely mirror human biases and exhibit common shifts in decision-making. Our audit raises important safety concerns: continued AI progress in both performance and human-machine value alignment may also increase security issues related to persuasion, manipulation, and deception.


SSRN and ArXiv

As a powerful and rapidly advancing dual-use technology, AI offers both immense benefits and worrisome risks. In response, governing bodies around the world are developing a range of regulatory AI laws and policies. This paper compares three distinct approaches taken by the EU, China and the US. Within the US, we explore AI regulation at both the federal and state level, with a focus on California's pending Senate Bill 1047. Each regulatory system reflects distinct cultural, political and economic perspectives. Each also highlights differing regional perspectives on regulatory risk-benefit tradeoffs, with divergent judgments on the balance between safety versus innovation and cooperation versus competition. Finally, differences between regulatory frameworks reflect contrastive stances in regards to trust in centralized authority versus trust in a more decentralized free market of self-interested stakeholders. Taken together, these varied approaches to AI innovation and regulation influence each other, the broader international community, and the future of AI regulation.


ACL EMNLP/CoNLL, Miami (November 12-16, 2024)

Stories are central for interpreting experiences, communicating and influencing each other via films, medical, media, and other narratives. Quantifying the similarity between stories has numerous applications including detecting IP infringement, detecting hallucinations, search/recommendation engines, and guiding human-AI collaborations. Despite this, traditional NLP text similarity metrics are limited to short text distance metrics like n-gram overlaps and embeddings. Larger texts require preprocessing with significant information loss through paraphrasing or multi-step decomposition. This paper introduces AIStorySimilarity, a novel benchmark to measure the semantic distance between long-text stories based on core structural elements drawn from narrative theory and script writing. Based on four narrative elements (characters, plot, setting, and themes) as well 31 sub-features within these, we use a SOTA LLM (gpt-3.5-turbo) to extract and evaluate the semantic similarity of a diverse set of major Hollywood movies. In addition, we compare human evaluation with story similarity scores computed three ways: extracting elements from film scripts before evaluation (Elements), directly evaluating entire scripts (Scripts), and extracting narrative elements from the parametric memory of SOTA LLMs without any provided scripts (GenAI). To the best of our knowledge, AIStorySimilarity is the first benchmark to measure long-text story similarity using a comprehensive approach to narrative theory. Code and data are available at https://github.com/jon-chun/AIStorySimilarity.


Frontiers in Computer Science

Affective artificial intelligence and multimodal sentiment analysis play critical roles in designing safe and effective human-computer interactions and are in diverse applications ranging from social chatbots to eldercare robots. However emotionally intelligent artificial intelligence can also manipulate, persuade, and otherwise compromise human autonomy. We face a constant stream of ever more capable models that can better understand nuanced, complex, and interrelated sentiments across different modalities including text, vision, and speech. This paper introduces MultiSentimentArcs, a combination of an open and extensible multimodal sentiment analysis framework, a challenging movie dataset, and a novel benchmark. This enables the quantitative and qualitative identification, comparison, and prioritization of conflicting sentiments commonly arising from different models and modalities. Diachronic multimodal sentiment analysis is especially challenging in film narratives where actors, directors, cinematographers and editors use dialog, characters, and other elements in contradiction with each other to accentuate dramatic tension. MultiSentimentArcs uses local open-source software models to democratize artificial intelligence. We demonstrate how a simple 2-step pipeline of specialized open-source software with a large multimodal model followed by a large language model can approximate video sentiment analysis of a commercial state-of-the-art Claude 3 Opus. To the best of our knowledge, MultiSentimentArcs is the first fully open-source diachronic multimodal sentiment analysis framework, dataset, and benchmark to enable automatic or human-in-the-loop exploration, analysis, and critique of multimodal sentiment analysis on long-form narratives. We demonstrate two novel coherence metrics and a methodology to identify, quantify, and explain real-world sentiment models and modalities. MultiSentimentArcs integrates artificial intelligence with traditional narrative studies and related fields like film, linguistic and cultural studies. It also contributes to eXplainable artificial intelligence and artificial intelligence safety by enhancing artificial intelligence transparency in surfacing emotional persuasion, manipulation, and deception techniques. Finally, it can filter noisy emotional input and prioritize information rich channels to build more performant real-world human computer interface applications in fields like e-learning and medicine. This research contributes to the field of interdisciplinary human-centered AI by giving non-artificial intelligence experts access to directly engage in analysis and critique of research around affective artificial intelligence and human-AI alignment. Code and non-copyrighted data will be available at https://github.com/jon-chun/multisentimentarcs.


Frontiers in Computer Science, August 12, 2024, Sec. Human-Media Interaction, Volume 6 - 2024

Machine translation metrics often fall short in capturing the challenges of literary translation in which translators play a creative role. Large Language Models (LLMs) like GPT4o and Mistral offer new approaches to assessing how well a translation mirrors the reading experience from one language to another. Our case study focuses on the first volume of Marcel Proust's "A la recherche du temps perdu," a work known for its lively translation debates. We use stylometry and emotional arc leveraging the newest multilingual generative AI models to evaluate loss in translation according to different translation theories. AI analysis reveals previously undertheorized aspects of translation. Notably, we uncover changes in authorial style and the evolution of sentiment language over time. Our study demonstrates that AI-driven approaches leveraging advanced LLMs yield new perspectives on literary translation assessment. These methods offer insight into the creative choices made by translators and open up new avenues for understanding the complexities of translating literary works.


ICML 2024, May 7-11, Vienna, Austria

"Risks and Opportunities of Open-Source Generative AI" (Long form version)

In the next few years, applications of Generative AI are expected to revolutionize a number of different areas, ranging from science & medicine to education. The potential for these seismic changes has triggered a lively debate about potential risks and resulted in calls for tighter regulation, in particular from some of the major tech companies who are leading in AI development. This regulation is likely to put at risk the budding field of open source Generative AI. We argue for the responsible open sourcing of generative AI models in the near and medium term. To set the stage, we first introduce an AI openness taxonomy system and apply it to 40 current large language models. We then outline differential benefits and risks of open versus closed source AI and present potential risk mitigation, ranging from best practices to calls for technical and scientific contributions. We hope that this report will add a much needed missing voice to the current public discourse on near to mid-term AI safety and other societal impact.


ArXiv.org (January 9, 2024)

With the rise of individual and collaborative networks of autonomous agents, AI is deployed in more key reasoning and decision-making roles. For this reason, ethics-based audits play a pivotal role in the rapidly growing fields of AI safety and regulation. This paper undertakes an ethics-based audit to probe the 8 leading commercial and open-source Large Language Models including GPT-4. We assess explicability and trustworthiness by a) establishing how well different models engage in moral reasoning and b) comparing normative values underlying models as ethical frameworks. We employ an experimental, evidence-based approach that challenges the models with ethical dilemmas in order to probe human-AI alignment. The ethical scenarios are designed to require a decision in which the particulars of the situation may or may not necessitate deviating from normative ethical principles. A sophisticated ethical framework was consistently elicited in one model, GPT-4. Nonetheless, troubling findings include underlying normative frameworks with clear bias towards particular cultural norms. Many models also exhibit disturbing authoritarian tendencies. Code is available at https://github.com/jon-chun/llm-sota-chatbots-ethics-based-audit.


Springer International Journal of Digital Humanities 5, 507–532 (2023). https://doi.org/10.1007/s42803-023-00069-8 (October 11, 2023)

The recent development of Transformers and large language models (LLMs) offer unique opportunities to work with natural language. They bring a degree of understanding and fluidity far surpassing previous language models, and they are rapidly progressing. They excel at representing and interpreting ideas and experiences that involve complex and subtle language and are therefore ideal for interdisciplinary human-centered AI research. This paper briefly surveys how XAI can be used to augment two interdisciplinary human-centered AI research areas relying on LLMs: (a) diachronic text sentiment analysis and (b) narrative generation. We also introduce a novel XAI greybox ensemble for diachronic sentiment analysis generalizable to any AI classification data points within a structured time series. Under human-in-the-loop supervision (HITL), this greybox ensemble combines the high performance of SOTA blackbox models like gpt-4-0613 with the interpretability, efficiency, and privacy-preserving nature of whitebox models. Two new local (EPC) and global (ECC) metrics enable multi-scale XAI at both the local and global levels. This greybox ensemble framework extends the SentimentArcs framework with OpenAI's latest GPT models, new metrics and a modified supervisory HITL workflow released as open source software at https://github.com/jon-chun/SentimentArcs-Greybox.


International Journal of Humanities and Arts Computing

This article outlines what a successful interdisciplinary human-centered artificial intelligence curriculum entails and why it is so critical now. Artificial intelligence is rapidly reshaping our world and is poised to exacerbate long-standing crises including (1) the crisis of higher education and the humanities, (2) the lack of diversity, equity and inclusion (DEI) in computer science and technology fields and (3) the wider social and economic crises facilitated by new technologies. We outline a number of ways in which an AI DH curriculum offers concrete and impactful responses to these many crises. AI DH yields meaningful new avenues of research for the humanities and the humanistic social sciences, and offers new ways that higher education can better prepare students for the world into which they graduate. DEI metrics show how an AI DH curriculum can engage students traditionally underserved by conventional STEM courses. Finally, AI DH educates all students for civic engagement in order to address both the social and economic impacts of emerging AI technologies. This article provides an overview of an AI DH curriculum, the motivating theory behind design decisions, and a detailed look into two sample courses.


Narrative 2023, March 1-4, Dallas, TX

Slides


"Augmenting Narrative Generation with Visual Imagery Using Integrated Prompt Engineering (ChatGPT, DALL-E 2)"

Narrative 2023, March 1-4, Dallas, TX

Conference Program


(video and links to generative AI resources), January 17, 2023, AI Colab, Kenyon College, Gambier, OH


The Helix Center, November 19, 2022, New York, NY


Narrative 30, no. 1 (2022): 104-113. doi:10.1353/nar.2022.0005.


ArXiv abs/2110.09454 (2021): n. page.

SOTA Transformer and DNN short text sentiment classifiers report over 97% accuracy on narrow domains like IMDB movie reviews. Real-world performance is significantly lower because traditional models overfit benchmarks and generalize poorly to different or more open domain texts. This paper introduces SentimentArcs, a new self-supervised time series sentiment analysis methodology that addresses the two main limitations of traditional supervised sentiment analysis: limited labeled training datasets and poor generalization. A large ensemble of diverse models provides a synthetic ground truth for self-supervised learning. Novel metrics jointly optimize an exhaustive search across every possible corpus:model combination. The joint optimization over both the corpus and model solves the generalization problem. Simple visualizations exploit the temporal structure in narratives so domain experts can quickly spot trends, identify key features, and note anomalies over hundreds of arcs and millions of data points. To our knowledge, this is the first self-supervised method for time series sentiment analysis and the largest survey directly comparing real-world model performance on long-form narratives.


In collaboration with Katherine Elkins, James Dennen (Denison University and Wexner Arts), Lauren Katz (Thymele Arts, LA)

100th anniversary of the premiere of "R.U.R.," by Czechoslovakian playwright Karel Čapek. "R.U.R." (for "Rossum's Universal Robots") opened on January 25, 1921, at the National Theater of Prague and marks the first use of the word "robot," coined by Čapek and derived from the Czech word for "forced labor." January 25, 2021


Journal of Cultural Analytics 5, no. 2 (2020): 17212.

Until recently the field of natural language generation relied upon formalized grammar systems, small-scale statistical models, and lengthy sets of heuristic rules. This older technology was fairly limited and brittle: it could remix language into word salad poems or chat with humans within narrowly defined topics. Recently, very large-scale statistical language models have dramatically advanced the field, and GPT-3 is just one example. It can internalize the rules of language without explicit programming or rules. Instead, much like a human child, GPT-3 learns language through repeated exposure, albeit on a much larger scale. Without explicit rules, it can sometimes fail at the simplest of linguistic tasks, but it can also excel at more difficult ones like imitating an author or waxing philosophical.


Narrative 2020 Conference (page 28), March 5-7, The Intercontinental Hotel, New Orleans


Cover of The Shapes of Stories by Katherine Elkins from Cambridge Elements Digital Literary Studies series

SentimentArcs is the open-source code for The Shapes of Stories by Katherine Elkins (Cambridge Press, August 2022)

Sentiment analysis has gained widespread adoption in many fields, but not—until now—in literary studies. Scholars have lacked a robust methodology that adapts the tool to the skills and questions central to literary scholars. Also lacking has been quantitative data to help the scholar choose between the many models. Which model is best for which narrative, and why? By comparing over three dozen models, including the latest Deep Learning AI, the author details how to choose the correct model—or set of models—depending on the unique affective fingerprint of a narrative. The author also demonstrates how to combine a clustered close reading of textual cruxes in order to interpret a narrative. By analyzing a diverse and cross-cultural range of texts in a series of case studies, the Element highlights new insights into the many shapes of stories.

Back to Top


Collaborations

Here is a sample list of collaborations across different institutions in academia, industry, and government:

  • US Whitehouse AI Safety Policy/NIST AI Safety Institute (NIST, Modern Language Assoc)
  • Collaborative AI Safety and Open Source Policy (Oxford, Berkeley, UVa, Berklee/MIT, etc.)
  • Tech Strategy and Investing (Panelists on monthly calls with BWG Strategy, NYC)
  • AI Alliance Essential AI Competencies Guide 2024 (Meta, IBM, Intel, US Chamber of Commerce, etc.)
  • AI and the Future of Higher Education (Notre Dame, CMU, UC Berkeley, Harvard Business School, LSU, St. Johns, etc.)
  • Computational Law, Ethics and eXplainable AI (upcoming: ACM, Northwestern Law School, IBM, Notre Dame)
  • Multimodal Affective AI/Emotions (Meta, NEH, EMNLP/CoNLL)
  • Computational Interdisciplinary human-centered AI (Dartmouth, U Tenn)
  • Literature, language, linguistics (MLA)
  • Narrative and Affective AI (Yale, Harvard, Narrative Conferences)
  • Diachronic Sentiment Analysis (Cambridge UP, ArXiv)
  • Economics and Law (IBM-Notre Dame)
  • NLP and AI Research (ICML, ACL, ACM)
  • Theatre and Film (Denison, NEH, LA Theatre)
  • Psychology and Cognitive Science: LatinX Summer Scholars (Ohio State University, Kenyon AI Colab)
  • Medicine / Federal Drug Administration (FDA, AFDO/RAPS)
  • Theoretical computer science (McGill, UPitt, CMU)
  • Physics and Engineering Simulations for Manufacturing (NAFEMS)
  • Internet of Things (IoT) real-time predictive maintenance (Kenyon AI Colab, Weyerhaeuser Paper)
  • Political Science and Public Policy (NIST, Meta, IEEE/EU)
  • Philosophy and Ethics (ArXiv, Montreal AI Ethics Inst)
  • Interdisciplinary Roundtables (The Helix Institute)
  • Judge for Startup Competitions and Entrepreneurship (UC Berkeley Graduate College of Engineering, Ohio State University)
  • International Startups: Latin American Supply Chain and Pan-Pacific Lifestyle Commerce (LA, São Paulo, Brazil, and Tokyo)
  • Computer and Network Security (USPTO, VPN appliance and secure packet relay patents, Symantec)
  • Electronic Medical Records (U of Iowa Hospitals and Clinics, UI Medical School, American Medical Informatics Journal)
  • Financial Information Systems, Asia Pacific (Bloomberg, McGraw-Hill Tokyo HQ)
  • Semiconductor Research and Competitive Analysis of Japanese Patents (SEMATECH, Lawrence Berkeley Labs)
  • Technical Japanese and Founder Hardware Test Lab (MIT Japan Program, DELL Computing)

Back to Top


Innovation in Higher Ed

I creatively apply industry best practices and state-of-the-art AI/ML techniques to high-impact interdisciplinary research. The combination of AI/ML, mathematics/statistics, and diverse domain expertise provides fresh insights and creates new paths of discovery. Our paper The Crisis of Artificial Intelligence: A New Digital Humanities Curriculum for Human-Centred AI, published in the International Journal of Humanities and Computing, outlines the challenges and opportunities AI presents to higher education, as well as the theory, structure, and goals of our human-centered AI approach that we hope can become a model for other institutions.

Our human-centered curriculum is designed to be collaborative at multiple levels and interdisciplinary while maintaining academic rigor. Students from virtually every department across campus create a unique classroom learning and research environment. Research from our AI Colab has contributed to numerous collaborative projects with institutions including Oxford, UC Berkeley, CMU, and Harvard Business School, with co-authors from departments spanning Computer Science, AI, Mathematics, Physics, Economics, and MBA programs. We focus on productive collaborations to operationalize solutions to the "big questions" central to the Liberal Arts experience. This includes partnerships with industry leaders (Meta, IBM), government agencies (NIST AI Safety Institute), and non-profits (AI Alliance, The Modern Language Association, US Chamber of Commerce, The Helix Center). Recent projects have addressed international AI regulation, researched ethical frameworks inherent in AI models, and reimagined education and skills development for an AI-integrated workforce (Guide to Essential Competencies for AI).

I am committed to bringing diverse voices to urgent debates surrounding technology's societal impact. Our interdisciplinary human-centered AI computing curriculum has successfully attracted a majority female enrollment (61%), non-STEM majors (91%), and historically underrepresented minorities (11% Hispanic, 13% Black) as of 2022. The program has grown to become one of the most popular courses on campus, with both faculty and student research generating significant academic impact through citations and institutional engagement worldwide.

Over the past decade, I have developed a human-first approach to teaching computation grounded in ML, AI, and Data Science, with real-world applications inseparable from ethics. This required bridging the C.P. Snow STEM and non-STEM divide, harmonizing academic specialization with practical interdisciplinary solutions, and building a new interdisciplinary human-centered AI computing curriculum without any additional institutional budget, headcount, or support resources.

In its first six years, our foundational course has become one of the most popular on campus. Faculty and student research has been published in leading journals and presented at major conferences, reaching tens of thousands of readers from prestigious universities and research centers globally. Both program founders have engaged with multiple organizations beyond our institution dedicated to AI ethics and CS education innovation.

My philosophical goal is to cultivate in students a technologically informed worldview grounded in universal humanistic values. It is imperative to equip all students, especially those underserved by traditional CS/AI programs, with a sound intellectual framework that synthesizes both sides of the C.P. Snow divide for both the individual and society at large.

Back to Top


Diversity from a Human-Centered AI Curriculum

Blackboard showing Fall 2022 class photo of IPHS 200 Programming Humanity students

UPDATE: Progress on UMR Diversity

Fall 2022 IPHS 200 Programming Humanity (estimate)

Category Count Percent
Male 41 53%
Female 36 47%
TOTAL 78 100%
  • 13% African-American (10)

Line graph showing progress on gender diversity in interdisciplinary human-centered AI curriculum since 2017-2018 academic year, reaching 61% female as of Spring 2022

Progress on Gender Diversity in interdisciplinary human-centered AI curriculum since the 2017-2018 academic year (61% female as of Spring 2022)

At Kenyon College, I co-founded the world's first human-centered AI curriculum. I am the sole technical advisor and the primary collaborative content creator. Over the last six years of teaching this curriculum, we have achieved the following milestones:

Research: Published research in top publications and conferences (Cambridge UP, Narrative, Journal of Cultural Analytics, etc.) with clear growth in citations.

Interdisciplinary Human-Centered AI Colab Research: Organically grew (no marketing/PR) to ~15k hits from top universities worldwide (#4 CMU, #5 Berkeley, #6 Stanford, #7 Columbia, #9 NYU, #16 Princeton, #22 Oxford, #23 MIT, #25 Cambridge, etc.)

Diversity:

  • Female: Grew from 18% to 61% between 2017-2021
  • Hispanic: participation rates are often at or above college averages
  • Black: 13% (Fall 2022 estimate above)
  • Non-STEM: Our classes are ~90% non-STEM from across nearly all departments, enfranchising many students who may otherwise feel alienated by traditional CS programs
  • 100% Pass rate (Quality of student work independently confirmed by success of their research archive at digital.kenyon.edu/dh)
  • 0% Drop rate

Enrollment: Experienced enrollment growth from 20 to 120 between 2017-2022 becoming one of the largest classes at Kenyon as an elective with no credit toward the traditional STEM computing major/minor

Budget: With no budget or antecedent, innovated from scratch a globally recognized interdisciplinary AI Colab research center and human-centered AI program. This includes no funds for hardware, software, cloud computing, support staff or other common expenses. This is achieved through continual strategic planning, careful curation and testing fully open-source, robust, best-of-breed and/or freely available resources informed by decades of experience in industry.

Our interdisciplinary AI DH research has been published in top presses, journals, and conferences. We have also mentored hundreds of ML/AI DH projects that synthesize Artificial Intelligence with literature, history, political science, art, dance, music, law, medicine, economics and more. Various sample AI DH projects are given at the bottom of this page.

Timeline


Kenyon College National Endowment for the Humanities Professorship history banner

Kenyon College's The National Endowment for the Humanities Professorship

Our AI research and AI Colab were collaboratively developed, and the curriculum is currently co-taught by a technology expert (Jon Chun) and an accomplished academic (Katherine Elkins). Both have broad experiences, publications, and interests transcending traditional domain boundaries. Support was provided with a 3-year National Endowment for the Humanities (NEH) appointment described here.


Katherine Elkins NEH Professorship announcement with photo and description of work

Collaborator Katherine Elkins work as Kenyon College's National Endowment for the Humanities Professorship


Bar chart showing over 90% non-STEM majors in the human-centered AI curriculum

A Humanity-First approach to interdisciplinary AI project-based learning consistently attracts over 90% non-STEM majors (Kenyon College Institutional Research)

Back to Top


Code, Products and Patents


Block diagram showing SentimentArcs architecture with data flow between components

Block Diagram for SentimentArcs Notebooks

Stories are everywhere. Here are a few examples of original research projects using SentimentArcs to extract and analyze narrative emotional arcs in:


Multimodal SentimentArcs line plot for Royal Wedding 1951 video with 10% simple moving average

Multimodal SentimentArcs: Royal Wedding (1951) Video 10% SMA Plot (2024)


Multimodal SentimentArcs line plot for Royal Wedding 1951 transcript with 10% simple moving average

Multimodal SentimentArcs: Royal Wedding (1951) Transcript 10% SMA Plot (2024)


Kernel density estimation plot comparing sentiment distributions for Royal Wedding 1951

Multimodal SentimentArcs: Royal Wedding (1951) KDE Plot (2024)

Back to Top


Kenyon Human-Centered AI Lab

Screenshot showing top 10 institutions reading Kenyon DH research in 2022 including CMU, Berkeley, and Stanford

Top 10 Institutions reading our AI DH Research in 2022 — digital.kenyon.edu/dh


Expanded list of leading institutions reading Kenyon AI DH research in 2022

Leading Institutions reading our AI DH Research in 2022 — digital.kenyon.edu/dh


Map showing Eurasian institutions accessing Kenyon DH research

Eurasian Institutions — digital.kenyon.edu/dh


Map showing institutions from the Americas accessing Kenyon DH research

Institutions from The Americas — digital.kenyon.edu/dh


Chart showing 149 countries accessing Kenyon DH research worldwide

Countries Worldwide — digital.kenyon.edu/dh


Chart showing 1,958 institutions accessing Kenyon DH research worldwide as of May 2023

Institutions Worldwide (May 2023) — digital.kenyon.edu/dh

Back to Top


Social Media

Twitter logo icon

@jonchun2000 — Main Social Media Account

Back to Top


Course Descriptions and Mentored Research

Whiteboard with brainstorming notes on literary analysis, financial forensics, and generative art prompts

Brainstorming to translate new theories into testable models for (a) Literary Analysis, (b) Financial Forensics and (c) the Latent Space of Generative Art Prompts.


IPHS 200 Programming Humanity (Introduced Fall 2017)

How do we think of humanity in the digital age? Can we program humanity into our technology by generating music, analyzing vast quantities of literary text, and producing great visual works of art? Or will humans be programmed through genetic engineering, predictive policing, manipulations of social media, and domestic surveillance? Are we firmly entrenched in a Posthumanist world, and is Transhumanism a future to which we should aspire? By the end of this course you will:

  • Have basic skills in Python, data visualization, data wrangling, and natural language processing.
  • Gain a foundational understanding of key aspects of our Age of Information from Dataism and Information Theory to Machine Learning and Artificial Intelligence.
  • Relate this emerging picture to conversations about Humanism, Posthumanism, and Transhumanism.
  • Understand how new computational tools both challenge and deepen traditional approaches to the Social Sciences, Humanities, and Fine Arts.
  • Understand emerging ethical debates surrounding key technologies poised to transform humanity.
  • Have a portfolio of project-based work demonstrating your computational skills with a final project of your choice.

Sample Course Projects:


IPHS 290 Cultural Analytics (Introduced Fall 2022)

IPHS 290 Cultural Analytics explores advanced computational methods for cultural analytics, focusing on the analysis of complex composite data types including geospatial, time series, and network graphs. Building upon foundational programming knowledge, students learn specialized techniques for collecting, transforming, and visualizing unstructured and structured data to investigate contemporary and historical cultural phenomena. Through engagement with interdisciplinary AI scholarship and hands-on technical implementation, students develop expertise in analyzing spatially, temporally, and semantically correlated data. This course prepares researchers to expand their analytical capabilities beyond traditional structured datasets to encompass the full spectrum of cultural data analysis.


"Why isn't AI a branch of mathematics?" Norvig and Russell ask in their seminal textbook Artificial Intelligence: A Modern Approach. "The first answer," they explain, "is that AI from the start embraced the idea of duplicating human faculties such as creativity, self-improvement, and language use" (18).

We explore AI from this humanist perspective by building a foundation to understand the creativity, self-improvement and language at the core of the discipline. Our approach blends the conceptual with hands-on exploration and application. We think through the humanist-centered theory and practice of artificial intelligence and machine learning from a liberal arts perspective that is highly interdisciplinary in nature. To that end, the course will touch on domains intertwined with AI like cognitive science and language and linguistics.

Textual analysis will receive our major attention and will provide an ideal foundation for students interested in computational approaches to literature. However we will also explore a wide range of textual applications that will be of interest to a broader range of students: textual analytics that help us explore disciplines like economics, politics, sociology, cultural studies, and psychology. The last weeks of the course will move to image, video and sound by building on previous skills acquired from textual analysis.

Sample Course Projects:


How is AI changing humanity? This course explores the realities of labor, knowledge work, and automation. Can the current generation of rapidly improving autonomous AI agents collaboratively (or independently) craft novels, social media political campaigns, commercial brand marketing strategies, financial analyst reports, software apps, music videos, or write original scientific research papers? For those with basic Python skills and advanced curiosity, get hands-on experience building your own personal digital twin and a peek into the future.

This is an upper-division course that offers an in-depth exploration of advanced AI concepts, focusing on interdisciplinary applications of large language models, AI information systems, and autonomous agents. Students will engage with a progressive curriculum, starting with a review of Python and a series of four hands-on projects, equipping them with the skills and knowledge necessary to innovate in the rapidly evolving field of artificial intelligence.

Sample Course Projects using Multi-Agent Networks to Automate Knowledge Workflow (New Course as of Fall 2024):


IPHS 484: Senior Seminar integrates academic research methodologies with entrepreneurial product development practices, guiding students through the creation of an innovative technical project that bridges scholarly contribution with real-world impact. Students develop comprehensive skills in research design, competitive analysis, technical architecture, product management, and performance metrics while producing a minimum viable product or proof of concept. Through intensive mentorship and peer collaboration, participants transform their academic foundations into tangible outcomes, whether aimed at scholarly publication or venture development. This capstone experience emphasizes independent project execution while providing structured support for students to synthesize their coursework into a launching pad for their post-graduate aspirations.

Sample Course Projects:


This advanced project-based independent study focuses on developing an end-to-end Industrial Internet of Things (IIoT) predictive maintenance system through collaboration with a local fabrication factory. Over fourteen weeks, students progress from foundational research through system implementation, combining academic theory with practical engineering. The curriculum covers IIoT architecture, time series analysis, machine learning for predictive modeling, real-time data streaming, cloud infrastructure, and user interface development. Through hands-on experience with both hardware and software components, students learn to design, implement, and deploy a complete IIoT solution while addressing real-world industrial challenges. The course emphasizes both theoretical understanding and practical implementation, culminating in a functional predictive maintenance system that serves actual business needs.

Back to Top


Course Summary

Slide showing the virtuous cycle and feedback between 3 models that guide interdisciplinary innovation

The virtuous cycle, feedback and tension between the 3 models that guide our interdisciplinary innovation

Integrated Program for Humane Studies (2017-)

Overview:

This upper-division course offers an in-depth exploration of advanced AI concepts, focusing on interdisciplinary applications of large language models, AI information systems, and autonomous agents. Over 15 weeks, students will engage with a progressive curriculum, starting with a review of Python and a series of four hands-on projects: (a) OpenAI API programming a GPT-based chatbot, (b) mechanistic interpretations of transformer internals using Huggingface Transformers, (c) Retrieval-Augmented Generation (RAG) using LangChain, and (d) simulations of autonomous multi-agent systems using AutoGen. The course includes four substantive subprojects and one final project, enabling students to apply theoretical knowledge to practical, real-world AI challenges. This course is designed to equip students with the skills and knowledge necessary to innovate in the rapidly evolving field of artificial intelligence, emphasizing both technical proficiency and ethical considerations. Introductory Python programming experience required.

Note: These 4 broad frontiers of AI research are rapidly evolving and based upon my AI research and industry consulting with Meta, IBM, the Whitehouse/NIST AI Safety Institute, etc. There is a constant flow of major new AI research, libraries, frameworks and startups nearly every week. Since this course will begin 9 months after this syllabus was written, expect updates to reflect the most recent in AI research breakthroughs, tooling, and industry best practices as of August 2024. Nonetheless, the class will be structured around these 4 broad and relatively consistent universal areas in AI.


Scientific Computing Mentored Projects (2020-)

  • SciComp Senior Seminar/Research
    • Noisy Time Series Filtering, Smoothing and Feature Detection
    • Narrative Metrics for NLG using LLM Transformers
    • Diachronic Sentiment Analysis Central Bank Speeches using SentimentArcs
  • SciComp Independent Study

Back to Top


Organizations

Principal Investigator (2024-) for the Modern Language Association

Announcement: For over 100 years, the MLA has become the principal organization of scholars in language and literature with over 25,000 members in over 100 countries. The MLA is joining more than 200 of the nation's leading artificial intelligence (AI) stakeholders to participate in a US Department of Commerce initiative to support the development and deployment of trustworthy and safe AI. Established by the Department of Commerce's National Institute of Standards and Technology (NIST) on February 8, 2024, the US AI Safety Institute Consortium (AISIC) brings together AI creators and users, academics, government and industry researchers, and civil society organizations to meet this mission. The MLA-sponsored team will be led by Katherine Elkins and Jon Chun at Kenyon College. The team will evaluate model capabilities with a special focus on linguistic edge cases and ethical frameworks.

The AISIC includes companies and organizations on the front lines of developing and using AI systems as well as the civil society and academic teams building the foundational understanding AI's potential to transform our society. Consortium members represent the nation's largest companies and innovative startups; creators of the world's most advanced AI systems and hardware; representatives of professions with deep engagement in AI's use today; state and local governments; and nonprofits. The consortium will also work with organizations from other nations in order to establish interoperable and effective safety around the world.


Cofounder (2023-)

About: Our mission is to facilitate efficient collaboration on interdisciplinary AI between individual researchers and domain experts separated by geographic, organizational, doctrinal, and legal divisions. We focus on human-centered AI topics like safety, bias, explainability, ethics, and policy grounded in careful experimentation and expert interpretation. Our goal is to enable fast, focused and flexible research funding and collaboration overlooked by traditional institutional research structures. We are a purely volunteer and wholly virtual non-profit corporation conducting human-centered AI research and education in the public interest.


The Helix Center, New York, NY

Executive Committee (2022-)

Round-Table: Living in Difficult Times, November 19, 2022

About: The original inspiration for interdisciplinary forums arose from the observations by our director, Dr. Edward Nersessian, of the constraints in both communication and creativity among scientists at professional meetings, fueled both by narrow specialization and the grant process, that with its demand for sharply defined investigation seemed, in fact, to be limiting curiosity and inquiry. This motivated him to form discussion groups drawing on multiple disciplines, the creative productivity of which inspired the formation of the Philoctetes Center for the Multidisciplinary Study of the Imagination.

Mission: The primary mission of The Helix Center is to draw together leaders from distinct spheres of knowledge in the arts, humanities, sciences, and technology for interdisciplinary roundtables, the unique format of which potentiates new ideas, new questions, and facilitates emergent creative qualities of mind less possible in conventional collaborations. Such a drawing together of leaders of various disciplines irrespective of their academic affiliation allows the Helix Center to function as a kind of university without walls. In addition, through audience attendance and its Q&A engagement with the roundtable participants, and live streamed and archived events, we aim to expand public understanding and appreciation of the sciences and technology, the arts and humanities.

Back to Top


Kenyon AI Colab logo with tilted text design

Kenyon AI Colab (Kenyon Human-Centered AI Colab)

Pinned Loading

  1. sentimentarcs_notebooks sentimentarcs_notebooks Public

    SentimentArcs: a large ensemble of dozens of sentiment analysis models to analyze emotion in text over time

    Jupyter Notebook 41 10

  2. iiot-time-series-prediction-system iiot-time-series-prediction-system Public

    An End-to-End Industrial IoT Time Series Prediction System

    Python 3

  3. cultural-analytics.github.io cultural-analytics.github.io Public

    Computational Cultural Analytics Course at Kenyon College

    1

  4. sentimentarcs_original sentimentarcs_original Public

    Sentiment Analysis for Narrative Arcs

    Jupyter Notebook

  5. sentiment-arcs sentiment-arcs Public

    Python

  6. multisentimentarcs multisentimentarcs Public

    A Novel Method to Visualize Multimodal AI Sentiment Arcs in Long-Form Narratives

    Python 4 3