• @bogdugg
    link
    English
    14
    edit-2
    9 months ago

    I played it - and if it was truly only made by two people is quite impressive - but it’s just alright. The world is very cool, and is structured around multiple levels of a tower each with their own language that you need to learn to progress. My main issue with the game is that the differences between these languages, and the puzzles built around them, aren’t particularly interesting or deep or varied. There are a few gems, but overall it’s much closer to a traditional adventure game than you might expect on first glance.

    That said, the art and world design are very cool.

    Edit: As an aside, it’s worth noting that the Steam Reviews metric is a tad misleading in a similar way to Rotten Tomatoes, in that it only gauges ratio of positive reviews, over what those reviews are actually saying. A universal consensus of a game being a 7/10 (if we assume 7/10 is positive) will appear “better” than a game where 99% of people believe it is a 10/10, but 1% think it sucks. It’s good at predicting whether you will like it, it is bad at predicting how much.

    • @[email protected]
      link
      fedilink
      English
      29 months ago

      Edit: As an aside, it’s worth noting that the Steam Reviews metric is a tad misleading in a similar way to Rotten Tomatoes, in that it only gauges ratio of positive reviews, over what those reviews are actually saying. A universal consensus of a game being a 7/10 (if we assume 7/10 is positive) will appear “better” than a game where 99% of people believe it is a 10/10, but 1% think it sucks. It’s good at predicting whether you will like it, it is bad at predicting how much.

      I’m not sure how relevant this is, since your described situation pretty much doesn’t happen. Like so many things in life, reviews are expected to follow a normal distribution. There are definitely counter-examples (e.g. shitstorms leading to massive downvote waves), but due to the large number of reviewers things should average out for normal cases.

      • @bogdugg
        link
        English
        29 months ago

        I suck at math, but if the mean is sufficiently over the “positive” threshold, and there’s a low standard deviation across reviews, wouldn’t this have the problem I describe? The more certain people are about the quality of good games, the less relevant the ratio becomes, which is perhaps the opposite of what you would want.

        • @[email protected]
          link
          fedilink
          English
          29 months ago

          I suck at math, but if the mean is sufficiently over the “positive” threshold, and there’s a low standard deviation across reviews, wouldn’t this have the problem I describe?

          Since Steam reviews are only positive or negative, not on a point scale, I’m not sure how this problem would come to pass. The distribution of reviews around the mean are expected to be similar for your described 10/10 game and the 7/10 game, and since the review system itself is only boolean in nature there is no distorted result.

          The more certain people are about the quality of good games, the less relevant the ratio becomes, which is perhaps the opposite of what you would want.

          Why does the ratio become less relevant the more certain people are about the quality of good games? Again, the review is only positive or negative, no actual review number assigned. In which cases do you expect the ratio to drift away from the actual useful information?

          • @bogdugg
            link
            English
            1
            edit-2
            9 months ago

            Why does the ratio become less relevant the more certain people are about the quality of good games? Again, the review is only positive or negative, no actual review number assigned. In which cases do you expect the ratio to drift away from the actual useful information?

            It’s because there’s no review number in combination with varying certainty that makes for bad information regarding judgment calls about quality. If people are certain the game is a 7/10, that could produce a better score than being less certain about an 8/10, because the wider distribution (less certainty) could put more reviews below the positive/negative threshold.

            The following reviews: 6/10, 6.5/10 , 7/10, 7.5/10, 8/10 will produce a 100% rating. More certain, less useful.

            The following reviews: 4/10, 6/10, 8/10, 10/10, 10/10 will produce an 80% rating. Less certain, more useful.

            It’s only consistent if you assume all games follow the same distribution, which is not how reviews work in my opinion. There are many websites which do surface score information, and they follow wildly different distribution patterns depending on many different factors.

            Again, it is useful for predicting whether you’ll like it, but bad for predicting how much you’ll like it.

            • @[email protected]
              link
              fedilink
              English
              19 months ago

              It’s because there’s no review number in combination with varying certainty that makes for bad information regarding judgment calls about quality. If people are certain the game is a 7/10, that could produce a better score than being less certain about an 8/10, because the wider distribution (less certainty) could put more reviews below the positive/negative threshold.

              But what do you mean with “more certain” and “less certain”? Again, Steam doesn’t have reviews beyond boolean values.

              The following reviews: 6/10, 6.5/10 , 7/10, 7.5/10, 8/10 will produce a 100% rating. More certain, less useful.

              And since Steam doesn’t have point-based reviews, the 100% rating is fully correct, as presumably each of these reviewers gave a positive review.

              The following reviews: 4/10, 6/10, 8/10, 10/10, 10/10 will produce an 80% rating. Less certain, more useful.

              How is it “less certain”?

              It’s only consistent if you assume all games follow the same distribution, which is not how reviews work in my opinion.

              Do you have statistical analyses that show this assumption to be wrong?

              • @bogdugg
                link
                English
                19 months ago

                My entire argument stems from the idea that you can ascertain quality from these ratings, which I am refuting. The rating is “correct” in that it is measuring something, and as long as people keep in mind what that something is, there is no problem. But this article, for example, uses the flood of positive reviews to make the case that it is one of the best of the year, which I believe is faulty reasoning.

                What I meant by certain is that the reviews are more clumped together (again if you had a score - even though it isn’t present presumably you could attach one to these reviews), so there’s more agreement among different people about the quality of the product. If you don’t agree that games can be more or less polarizing, you won’t agree with this point unless I can back it up with data which I’m not going to spend time doing. You could go through Rotten Tomatoes and compare Critic Score with RT Score because they surface both those values and see how closely they track on different parts of the spectrum.