Racist or Just Biased? It’s Complicated.

Note: This is a continuation of the shared blogging of Warr, Mishra, and Oster. In this post, Melissa wrote the first draft to which Punya and Nicole added substantial revisions and edits. “Science” is social. We build on each other’s ideas. We critique each other’s ideas. And that’s how these ideas get refined; and, hopefully, […]

Read more
GenAI is Racist. Period.
GenAI is Racist, period.

Note 1: We have written a follow-up post that delves deeper into the “racist” vs “biased” responses- Racist or Just Biased? It’s Complicated Note 2: The shared blogging with Punya Mishra and Nicole Oster continues. Punya crafted the student essay and I generated and analyzed the data. Punya wrote the first draft which was then edited […]

Read more
Should LLMs Have “More Complex Predictive Capabilities”? 🤔Implications for Personalized Learning

Today ChatGPT4o (the o stands for “omni” apparently 🤷🏼‍♀️) was helping me summarize some research I was working on. I gave it some slides of what I was analyzing and, after a few back-and-forths, it told me this: Overall, these differences highlight an evolution in the significance and influence of demographic and socioeconomic variables between […]

Read more
ChatGPT Doesn’t Have a User’s Manual. Let’s Not Create One.

Note: This is the next post in the shared blogging experiment with Punya Mishra and Nicole Oster. This time we question what and how we should be teaching about generative AI. The core idea and first draft came from me, to which Punya and Nicole added revisions and edits. The final version emerged through a […]

Read more
Not for the Truth of the Matter

Note: Featured image made with Adobe Firefly 3. I’ve written a fair number of posts lately where I’ve explored my experiences with large language models, like ChatGPT, and questioned whether what was happening was a “new” type of learning or simply an amplified or enhanced process that is basically the same as the other learning […]

Read more