๐Python & Pizza (Fall 2025)
Event 1

This hands-on session is all about exploring how we can use Python and Large Language Models (LLMs) to analyze online discussions about ChatGPT. Youโll get to see how data from Reddit can be collected, analyzed, and turned into meaningful visual insights comparing the perspectives of professors and students.
๐ Access the Workshop Materials
I have prepared a Google Drive folder with everything you need to follow along:
- ๐ Colab Notebook โ ready-to-run code for scraping, analyzing through GPT, and visualizing Reddit posts.
- ๐งพ Sample Datasets โ posts from professorsโ and studentsโ related subreddits (about ChatGPT).
- ๐ Output Examples โ labeled CSV files and example plots.
๐ Click here to access the Google Drive folder
๐ง What Youโll Do
- Use Python in Colab (no installation required).
- Log into your gmail account
- Copy the Google drive folder I shared above in your Google drive before running
- you can download, extract, and upload to your drive or
- Just copy – this will add a “Copy of” tag in the name of the file which you will need to remove
- Label Reddit posts with sentiment, stance, and themes using GPT-4o-mini. (Already done for you)
- Run the cells to create visualizations like:
- Bar charts of sentiment and stance
- Top themes chart for professors and students
- Word clouds of themes
- Theme co-occurrence networks etc.
- When you go back home manually evaluate the posts and the labels to check how gpt-4o-mini performed on the data
- Zoom Link (in case you want to remotely join): https://suny-ow-edu.zoom.us/j/7222550367
๐ฌ Questions?
For more details about the Data Science program at SUNY Old Westbury, or to connect about future events, or bug reporting for the given codes, contact Dr. Shebuti Rayana at rayanas@oldwestbury.edu
โจ Grab some pizza, open the notebook, and start discovering insights from the online conversation about chatGPT using GPT ;)!
Event 2 – Thursday, October 30th
from 5:30 PM – 7:00 PM

๐ Access the Workshop Materials
- ๐ Colab Notebook โ ready-to-run code loading, analyzing data, and creating plots.
- ๐งพ Sample Datasets โ multiple climate datasets from NASA, and Berkley.
- ๐ Output โ climate plots.
๐ Click here to access the Google Drive
๐ง What Youโll Do
- Use Python in Colab (no installation required).
- Log into your gmail account
- Copy the Google drive folder I shared above in your Google drive before running
- you can download, extract, and upload the flies to your drive or
- Just copy – this will add a “Copy of” tag in the name of the file which you will need to remove
- If you do not plan to run the cells to regenerate the charts you can only copy the notebook .ipynb file
- Analysis and visualizations are already done for you
- The per-country temperature data in the temperature-data folder may be a little large – if you do not regenerate the globe with animation – you do not need to copy this folder to your drive
- Your task is to complete building the data story that I initiated in my presentation, by
- Following the storytelling arc – add the falling action and resolution to the story – you can make 1-5 slides for that. Hint: find a dataset for sea level increase, etc. Or you can find some charts/graphs online to add to the story.
- Zoom Link (in case you want to remotely join): https://suny-ow-edu.zoom.us/j/7222550367
Event 3 – Wednesday, December 3rd from 1:30 PM – 2:30 PM