Reproducibility is a bedrock of scientific progress. However, it’s remarkably difficult to get reproducible results out of large language models. For example, you might observe that asking ChatGPT the same question multiple times provides different results. This by itself is not surprising, since getting a result from a language model involves “sampling”, a process that converts the language model’s output into a probability distribution and probabilistically selects a token. What mig...| Thinking Machines Lab
The damage of grade inflation has largely been done as there’s no hope of returning to a grading scheme where a C is average. But we can still resolve the growing threat of grade compression before it wreaks further havoc on Harvard’s education. Maybe it’s time for another “Committee on Raising the Standard” — 128 years later.| www.thecrimson.com