• _stranger_@lemmy.world
    link
    fedilink
    arrow-up
    43
    ·
    edit-2
    3 hours ago

    You’re giving this person a lot of credit. It’s probably all in the same table and this idiot is probably doing something like a for-loop over an integer range (the length of the table) where it pulls the entire table down every iteration of the loop, dumps it to a local file, and then uses plain text search or some really bad regex’s to find the data they’re looking for.

    • indepndnt@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      2 hours ago

      I think you’re still giving them too much credit with the for loop and regex and everything. I’m thinking they exported something to Excel, got 60k rows, then tried to add a lookup formula to them. Since you know, they don’t use SQL. I’ve done ridiculous things like that in Excel, and it can get so busy that it slows down your whole computer, which I can imagine someone could interpret as their “hard drive overheating”.

    • morbidcactus@lemmy.ca
      link
      fedilink
      arrow-up
      18
      ·
      3 hours ago

      Considering that is nearly exactly some of the answers I’ve received during the technical part of interviews for jr data eng, you’re probably not far off.

      Shit I’ve seen solutions done up that look like that, fighting the optimiser every step (amongst other things)

    • makingStuffForFun@lemmy.ml
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      2 hours ago

      I have to admit I still have some legacy code that does that.

      Then I found pandas. Life changed for the better.

      Now I have lots if old code that I’ll update, “one day”.

      However, even my old code, terrible as it is, does not overheat anything, and can process massively larger sets of data than 60,000 rows without any issue except poor efficiency.