Currents 088: Melanie Mitchell on AI Measurement and Understanding
Jim talks with Melanie Mitchell about her critique of applying standardized exams to LLMs and the debate over understanding in AI. They discuss ChatGPT and GPT-4’s performance on standardized exams, questioning the underlying assumptions, OpenAI’s lack of transparency, soon-to-be-released open-source LLMs, prompt engineering, making GPT its own skyhook to reduce hallucinations, the number of parameters in […]
Currents 088: Melanie Mitchell on AI Measurement and Understanding Read More »







