• 6 Posts
  • 272 Comments
Joined 3 years ago
cake
Cake day: June 18th, 2023

help-circle

  • I think you’re misunderstanding a bit: What they did was primitive compared to what we do today, and the comparison is fair since what they did was the foundation for what we do today.

    Much like we could call a vacuum tube computer that filled a room in the 50’s a “primitive computer”. It’s not downplaying the fact that that computer was (and is) an absolute feat of science and engineering that very few people alive could replicate. It’s bringing focus to the fact that, despite being comparatively simple relative to a modern phone, it’s what made development of modern computers possible. It’s quite literally a “primitive computer”.

    In the same way, this procedure was an early predecessor to modern surgery, so it’s quite literally “primitive surgery”.


  • Oh, I definitely agree that meaningful whitespace can be a pain, and I’m not a very big fan in general (although I prefer meaningful whitespace like in Python to terribly indented code with regards to readability). I guess my point was just that if you’re having a lot of failures due to incorrect indentation, it sounds like a systemic issue somewhere. While meaningful indentation can be annoying, I think it sounds like symptom of something more if it’s responsible for most of your production failures.

    I think the bottom line for me is that if a config file regularly causes errors because of incorrect indentation, it should probably be refactored, since it’s clearly not easy enough to read that those errors are caught.


  • To me, that makes it sound like you’re writing too much and too complex yaml files manually, and/or that you don’t have good enough CI to catch invalid configurations. Unless, of course, you have very few prod failures overall, and the few that happen are due to yaml indentation, which I still think is a bit weird, since an invalid config caused by incorrect indentation should ideally be caught at compile time (if you’re generating code from the yaml) or by some linter or something (if you’re using it for config).


  • I’ll agree that significant whitespace can be a PITA (one of the reasons I prefer json over yaml), but at the same time I think improper or lacking indentation hurts readability even more than significant whitespace. Toml basically encourages a completely flat structure, where objects and sub-objects are defined all over the place. At that point, I much prefer an enforced structure with whitespace.













  • I’m a researcher myself, so I feel like I can weigh in on the “reproducibility crisis”. There are several facets to it: One is of course money, but that’s not just related to corporately funded research. Good luck finding or building an independent lab capable of reproducing the results at CERN. It basically boils down to the fact that some (a lot of) research is insanely expensive to do. This primarily applies to experiments and to some degree to computationally expensive stuff.

    Another side is related to interest. Your average researcher is fired up by the thought of being the first person to discover and publish something no one has seen before. It’s just not as fun to reproduce something someone else has already done. Even if you do, you’re likely to try to improve on it somehow, which means the results may change without directly invalidating the old results. It can be hard work to write a good paper, so if you don’t feel your results are novel enough that they’re worth the effort (because they’re basically just equivalent to previously published values) you might not bother to put in the effort to publish them.

    Finally, even without direct reproduction of previously published results, science has a way asymptotically approaching some kind of truth. When I develop and publish something, I’m building on dozens of previously published works. If what they did was plain wrong, then my models would also be liable to fail. I’ve had cases where we’ve improved on previously published work, not because we tried to reproduce it, but because we tried to build on their results, and found out that their results didn’t make sense. That kind of thing is fairly common, but not reported as a “reproduction study”.

    There’s also review articles that, while they don’t do any reproduction themselves, collect and compare a bunch of comparable work. They usually have some conclusions regarding what results appear trustworthy, and what appear to be erroneous.