Yes theres absolutely tools to convert mc worlds to minetest/luanti worlds. using these tools probably requires some willingness to run terminal based shell commands though.
Yes theres absolutely tools to convert mc worlds to minetest/luanti worlds. using these tools probably requires some willingness to run terminal based shell commands though.
Have you been keeping tabs on the mineclone projects these past few years? Mineclonia and voxelibre are both excellent Foss alternatives. Most people would struggle to be able to identify MCLA with a MC texture pack as any different from real deal.


Something that hasn’t been mentioned much in discussions about Anubis is that it has a graded tier system of how sketchy a client is and changing the kind of challenge based on a a weighted priority system.
The default bot policies it comes with has it so squeaky clean regular clients are passed through, then only slightly weighted clients/IPs get the metarefresh, then its when you get to moderate-suspicion level that JavaScript Proof of Work kicks. The bot policy and weight triggers for these levels, challenge action, and duration of clients validity are all configurable.
It seems to me that the sites who heavy hand the proof of work for every client with validity that only last every 5 minutes are the ones who are giving Anubis a bad wrap. The default bot policy settings Anubis comes with dont trigger PoW on the regular Firefox android clients ive tried including hardened ironfox. meanwhile other sites show the finger wag every connection no matter what.
Its understandable why some choose strict policies but they give the impression this is the only way it should be done which Is overkill. I’m glad theres config options to mitigate impact normal user experience.


Theres a compute option that doesnt require javascript. The responsibility lays on site owners to properly configure IMO, though you can make the argument its not default I guess.
https://anubis.techaro.lol/docs/admin/configuration/challenges/metarefresh
The metarefresh challenge sends a browser a much simpler challenge that makes it refresh the page after a set period of time. This enables clients to pass challenges without executing JavaScript.
To use it in your Anubis configuration:
# Generic catchall rule
- name: generic-browser
user_agent_regex: >-
Mozilla|Opera
action: CHALLENGE
challenge:
difficulty: 1 # Number of seconds to wait before refreshing the page
algorithm: metarefresh # Specify a non-JS challenge method
This is not enabled by default while this method is tested and its false positive rate is ascertained. Many modern scrapers use headless Google Chrome, so this will have a much higher false positive rate.


You know the thing is that they know the character is a problem/annoyance, thats how they grease the wheel on selling subscription access to a commecial version with different branding.
https://anubis.techaro.lol/docs/admin/botstopper/
If you want to use Anubis but organizational policies prevent you from using the branding that the open source project ships, we offer a commercial version of Anubis named BotStopper. BotStopper builds off of the open source core of Anubis and offers organizations more control over the branding, including but not limited to:
In the near future this will expand to:
In order to sign up for BotStopper, please do one of the following:
I have to respect the play tbh its clever. Absolutely the kind of greasy shit play that Julian from the trailer park boys would do if he were an open source developer.



Hi, hope you don’t mind me giving my two cents.
Local models are at their m9st useful in daily life when they scrape data from a reliable factual database or from the internet and then present/discuss that data to you through natural language conversation.
Think about searching for things on the internet now a days. Every search provider stuffs ads in top results and intentionally ofsucates the links your looking for especially if its a no-no term like pirating torrent sites.
Local llms can act as an advanced generalized RSS reader that automatically fetches articles and sources, send STEM based queries to wolfram alpha llm api and retrieve answers, fetch the weather directly from openweatherAPI, retrieve definitions and meanings from local dictionary, retrieve Wikipedia article pages from a local kiwix server, search ArXiv directly for prior searching. One of Claude’s big selling points is the research mode toolcall that scrapes hundreds of sites to collect up to date data on the thing your researching and presenting its finins in a neat structured way with cited sources. It does in minutes what would traditionally take a human hours or days of manual googling.
There are genuine uses for llms if your a nerdy computer homelab type of person familiar databases, data handling and can code up/integrate some basic api pipelines. The main challenge is selling these kinds of functions in an easy to understand and use way for the tech illiterate who already think bad of llms and similar due to generative slop. A positive future for llms integrated into Firefox would be something trained to fetch from your favorite sources and sift out the crap based on your preferences/keywords. More sites would have APIs for direct scraping and the key adding process would be one click button
I’m only a hobbyist no skin in the game but as an engineer first programmer second I feel like how you design the architecture of your logic and how you choose to prioritize computational optimization or feature implementation have a much bigger impact than language. Obviously theres a time and place like using luajit instead of lua if you really need that speed boost, but in big complex programs I’m willing to bet lazy programmers skip optimization steps and copy paste thousands of lines to ship quickly regardless how performance optimized the underlying language is.
I make a module that does a specific thing for version 1, after initial bug patches and maybe a feature addition theres usually an epithany that the logic can be generalized into an api type registry structure that allows the code to be reused and configured as needed on per item basis. Along the way you do stuff like ask what the not expensive calls are, add accumulator timers to reduce unneeded CPU usage , ensure early returns are up top when it makes sense, choosing to render a cleverly crafted sprite instead of a full 3d entity.
A conditional if then logic loop Is a logic loop no matter what Turing complete language its written in. What really matters is how you use the loop not which syntax its written in.