• kmerroll 6 days ago

    Aside from the Knowledge Graph buzzword, isn't this exactly the same idea as Tim Berners-Lee and the Semantic Web back in 2001? - web of resources, not pages - ontology-based schema - RDF based encoding - URI (IRI) resource identifiers - Automated agents and reasoning (DL) support

    Considering the ensuing reception and general avoidance of the semantic web outside academic papers, I guess no one wants to talk about it.

    • shubb 6 days ago

      And there are related standards like HATEOAS for state full behaviour right?

      But this isn't about a new way of presenting information to rival sparql or something.

      This is a technical report about a guy who wrote a slightly advanced crawler bot that discovers the behaviour of a modern Web application and can maybe be used to generate automated tests.

      It has a lot in common with that post about a new yc company using llms to generate integrations for existing undocumented Web applications.

      • euroderf 6 days ago

        The paper seems kind of dismissive:

        > Knowledge graph-based systems, such as Squirrel [8], have been proposed to crawl the semantic web and represent data in structured formats. However, these approaches are typically limited to RDF-based web data and do not capture the dynamic, user-driven interactions seen in modern web applications.

        • zcw100 4 days ago

          How is this dismissive? It's a fairly straightforward statement of fact. The only way I could possibly read it as dismissive is if you interpreted it like, "These approaches are typically limited to [old, boring ass,] RDF-based web data and do not capture the [rich, luscious] dynamic, user driven interactions seen in [beautiful] modern web applications"

          • euroderf 4 days ago

            "Dismissive" in the sense of "let's dismiss it for the remainder of this paper, because it does not apply here".

      • miningape 6 days ago

        Isn't this just going back to the 90s web? Before all the javascript and interactivity craziness was used. We can argue all day about how bad js is from a development perspective but I definitely like the interactivity it's brought to the web.

        • PittleyDunkin 6 days ago

          > but I definitely like the interactivity it's brought to the web.

          I don't!

          EDIT: to be earnest, I sincerely wish the web as it stands had chosen some serious distinction between "documents", "documents-with-scripting-but-no-web-requests", and "interactive applications with remote state". As it stands the last option trivially passes for the first two options.

          • wwweston 6 days ago

            JS can do some cool things. Sometimes it's necessary. Sometimes it's the most expedient way.

            On the other hand, we've exported high enough complexity and code payload sizes to the front-end and seen bandwidth rise enough that it's possible for some full page reloads to beat responsiveness/interactivity of many front ends at this point. But if one doesn't want to do that, there's techniques to let JS act primarily in the controller role while letting the back end take care of model/view concerns that, exchanging HTML as an adequate data serialization and display/UI description format.

            And, just as importantly, as a hypertext and serialization format, it provides opportunity for machine discovery and description of available resources. The ones we're most used to are search engines, but we're obviously in an era where we're discovering what else is possible.

            Possible ML models can also work with JS messes, but since they're less legible for people, there's less likely a good corpus correlating accurate descriptions with code in the wild. Legibility of what you feed the client has benefits beyond executablity, something that people who grew up with the early web were apparently prepared to understand.

          • sroussey 6 days ago

            This begs for the reverse-turning state graphs back into UI.