tupd 0.6 - small bugfix

S_S@lemy.lol to Selfhosted@lemmy.world – 10 points –

the server component had a small bug, empty checking the wrong variable before building a list of allowed zones

when using a config without any Fqdns defined this would result in the server refusing the client access to tunnel anything if any zones where about to be used


tup proxies services on the local network to a remote gateway, all traffic between the remote server and the service on the local network is sent through a wireguard tunnel

think of tup as an open source and self-hosted alternative to ngrok and cloudflare tunnel

tupd (the server) can be found at: https://drive.proton.me/urls/GEJM1HT0DW#aOop4p7zxaPA

the tup client can be found at: https://drive.proton.me/urls/63SE9PW020#GFzZrprg9wjZ


I also noticed all file extensions were not inspectable directly in the drive (even though everything is only text files), I apologize for that, I believe transparancy is a very important key factor

I've complemented with .diff files generated with diff from GNU diffutils, there are 'full' diff files for both tup and tupd (ending with _full.diff), and there is also a diff file for only the changes between tupd-0.5 and tupd-0.6 (tupd-0.6.diff)

the 'full' diff files can also be applied to an empty directory with GNU patch like this:

mkdir tupd-0.6
patch --directory=tupd-0.6/ --strip=1 < tupd-0.6_full.diff

Since my project is not uploaded by me to any git service many people didn't look on how it can be used so I want to give a few examples of the client, more explanations and examples can be found in the README.md and EXAMPLE.md of the client

Syntax: tup [-zone <zone>] [@][host]:[#]<port>

Examples:

tup :8080 this would proxy http://127.0.0.1:8080 onto a random subdomain on default zone, for example: https://xyz123.zone.domain.tld

tup 192.168.1.11:8080 this would proxy http://192.168.1.11:8080 onto a random subdomain on default zone

Syntax: tup -fqdn <domain> [@][host]:[#][@]<port>

Examples:

tup -fqdn sub.domain.tld :8080 this would proxy http://127.0.0.1:8080 directly onto https://sub.domain.tld

tup -fqdn sub.domain.tld 192.168.1.11:@8443 this would proxy https://192.168.1.11:8443 directly onto https://sub.domain.tld, skipping caddy and its tls termination on the server, same as a raw tcp proxy / sni proxy

Syntax: tup -udp|-tcp [rport:][@][host]:<port>

Examples:

tup -udp :27015 this would proxy udp://127.0.0.1:27015 onto a random UDP port on the server

tup -udp 27016:27015 this would proxy udp://127.0.0.1:27015 onto UDP port 27016 on the server

tup -udp 27016:192.168.1.11:27015 this would proxy udp://192.168.1.11:27015 onto UDP port 27016 on the server

tup -tcp :3306 this would proxy tcp://127.0.0.1:3306 onto a random TCP port on the server

I also want to clarify that the code is released with the Unlicense template, dedicating my software to the public domain

18

Gotta be honest, downloading security related software from a random drive is sending off sketchy vibes. Fundamentally, it's no different than a random untrusted git repo. But, I really would suggest using some source control rather than trying to roll your own with diff archives.

Likewise, I would also suggest adding in some unit and functional tests. Not only would it help maintain software quality, but also build confidence in other folks using the software you are releasing.

I totally agree it is no different than a random untrusted git repo, so I believe no additional trust is gained if I uploaded it to any of them

I think version control this way is totally fine, every commit in the linux kernel is mailed as a text diff on the different mailing lists

As of trusting this or any security related software I believe you have to ultimately read and understand the software you are using, or someone you trust has to do it, I can't do that, I can only answer questions as they arrive

I also agree unit tests are probably a good idea for those reasons as well, I don't have any right now but I'm open to do them some time or receive patches with them

I like your feedback, thanks for it

Git was literally written by Linus to manage the source of the kernel. Sure patches are proposed via mailing list, but the actual source is hosted and managed via git. It is literally the gold standard, and source control is a foundational piece of software development. Same with not just unit tests, but functional testing too. You absolutely should not be putting off testing.

I've done a lot of testing, not skipping that, writing automated tests are a whole different thing however, it is not as straight forward and is very often skipped for a large amount of projects to be honest

Git was made to handle the sheer amount of commits and people contributing to the linux kernel, the first versions of linux is just Linus uploading the code to a FTP, git is just a tool for Linus to patch his local git tree in a fast way with all the patches he gets from different channels and manage a large public repository

Unlike Linus I'm not planning to be of control of a public development process for my software, so a VCS doesn't make much sense in my opinion

Before git it was far from standard to use a source control system on small projects that weren't about to be a public development process anyway, while it is a gold standard for source control today, I don't think one have to use source control on every software project, like it used to be

I like the idea, but may I ask why you don't use a version control system like Git (or anything else that fits your needs)?

Thanks

I just don't think I need to use one on every one of my projects, it didn't add any (or enough) value to me for this project for bringing in another tool into the development or release process

For me this isn't controversial to skip either, I'm using (and even contributing some to) software all the time with just using folder archives without needing to use any repo tools, historically a lot of software development haven't need it as well

Github and Gitlab are free, and both even allow private repos for free at this point. Git is practically one of the first tools I install on a dev machine. Likewise, git is the defacto means of package management in golang. It's so built in that module names are repo URLs.

Yes I know it is built in into the package management in golang and other languages, I'm pulling things into my project that way

I know it is free also and that I can have private repos, it is one of the first tools I also install on dev machines

But I don't see how that means I always have to use git or any other VCS for every project I make, it is a good tool, but why do I have to use it if it doesn't help me?

You don't have to. Absolutely not.

But: As a potential user it provides some additional features your solution lacks. I can easily fork or clone your repo and change things if I need to. If I think it benefits the project I can easily offer these changes back to you, if I don't I can still profit from future development on your side and incorporate my changes into it. I can very easily check what has changed between two versions without relying (trusting) your changelogs or performing a manual diff.

But most importantly it is a matter of trust. Not so much trust in your intentions and the possibility of malicious code (Git won't prevent that), but it obfuscates your code unnecessarily making it harder to continue if you at some point decide to stop maintaining it or even detect vulnerabilities as it is not easily accessible without knowing where to look for it.

Sure it might creates some hurdles for potential contributors, and cause some discomfort for users

For the user case I believe downloading the archive or even the diff text and applying it is quite straight forward, there isn't any more steps in that than if it was in a repo, it is another way of pulling the code, I'm pretty used to handle software this way anyway, but most people aren't, I even doubt people would have pulled the SVN URL if I had released it on SourceForge for example, they would have downloaded the source archive from the website, so I don't believe that the fact that I don't upload it to a git/svn service makes it much harder for a user

It is a lot more work to setup a git repo, create an account somewhere, manage ssh keys, setting up configs, making commits, pushing code etc for me

Offers changes back to me is more a me problem, if you create a git repo and send me the patches or URL I can figure the rest out, anyone doing that can still benefit from my diff files as they can be applied to a git repository as well, or they could create two repositories (or branches), one with just my changes and one with theirs mixed in, and get full three way merge as well

For the easily checking what has changed it is quite the same thing, there is an extra step unfortunately

For the relying and trusting me all I can offer is the transparency, I don't believe many people would notice when a project get forced pushed either, some developers might use the same directory every time to make a pull, and then it would tell them there's a mismatch, but it is a legit concern nonetheless

I think using git solves that a bit, you would notice if something was forced pushed, as the auto-updating would fail with an error, but applying patches would also fail the same way, but that is of course a manual operation, having it in a SVN repo feels like it would produce the same problem also

SVN is probably as much of obfuscating as these files because git is what everyone uses, so it is more a problem of not using git than not using a (D)VCS I believe

Yea finding it is hard, there are a lot of other benefits with centralized social development platforms, but there was also a time before all that started, I think this community in particular is about not giving up to centralized platforms

If I ever stop the development, die, go rouge, get hacked or anything else I hope the community figures that out on their own and solves the problem then, anyone who wants should download the text diff files and/or the tar.gz files and take a few different checksums of them so they do not change maliciously at least

I think you have realized that every comment here was about your decision not to use Git. I don't think there is much more to say about this...

Yes unfortunately most of the comments on my last post was also about me not using git and git platforms, seems very controversial.

I totally understand it, I've also wanted to have everything git and social development platform before when it was new

This isn't about social platforms or using the newest-hottest tech. It's about following industry standard practices. You act like source control is such a pain in the ass and that it's some huge burden. And that I just don't understand. Getting started with git is so simple, and setting up an account with a repo host is a one time thing. I find it hard to believe that you don't already have ssh keys set up too. What I find more controversial and concerning is your ho-hum opinion on automated testing, and your belief that "most software doesn't do it". You're writing software that you expect people to not only run on their infra, but also expose to the public internet. Not only that, but it also needs to protect the traffic between the server on public infra and client on private infra. There is a much higher expectation of good practices being in place. And it is clear that you are willingly disregarding basic industry standard practices.

It is obviously about that since no good arguments are really made for anything else. There is just a constant nonsense talk why to use git and git platforms. Developers seems to have got stuck in some mindset that everything needs to version controlled all the time, if you make a change or at least a couple you gotta have to make a git commit, or otherwise most of the benefits are lost anyway, but of course I don't really know how much of developers you really are. I don't care what the industry standard is for public development, I'm not making public development, I'm only making a public release here. Are you really saying that using a VCS isn't a pain in the ass? I do it because I most and when someone pays me to do it. Everyone is acting like it is just a natural piece in the workflow of any software project, I strongly disagree. It is in the way and causing extra job in every step on the way (or is everyone using a auto-commiter or a full blown CI/CD system for everything?). But that is the point when you develop something in group or for money, it has strong benefits then. Here I have time to develop it linearly and in my own pace and do all the testing I need. I don't need git as a CTRL+Z tool or as a unfinished feature branch tool. My text editor alone got me covered, I've trusted it with my life for over 10 years. I couldn't be happier and more productive with it than I already am. It might be simple to use git, but it is still extra, and in this case unnecessary extra. It literally doesn't give me anything. Why would I also setup a git provider account, even if it is one time, ssh keys are made in a split second but you got to manage them as well, add them to the provider website and take them with you where you want to develop. I got all that setup on private instances, I fully manage myself, I know exactly how everything of that works.

It is a fact that most software skip unit tests and functional tests, I don't believe it is good and neither do I brag about that I don't have that yet. This project doesn't need to be exposed to the public internet at all. You can run them perfectly fine in a own VLAN or something. I'm not saying anyone should run anything they aren't fully trusting. I would much rather be discussing design decisions or security concerns than why I'm not using git or have any unit tests yet. I know exactly how sensitive it is for software to run in someones infrastructure, and tools that could alter and manage that, and protect the traffic between the node and the endpoint. But then give something real critical of that and refer to the code or design and not just good practices. I know they are good and useful, but it is not everything. What matters is the code in the end. Don't run if you can't trust it because its lacking automated testing then. I don't care if I'm the only one ever using it, I created it for myself and it is solving a whole lof of issues no other tool or solution comes close to. I'm sharing it if anyone else is looking for something like this and maybe wants to just take a look on how I solved it when developing their own tools. I'm more than happy to help with setting it up or answering any questions related to it or the things it does

Why not publish on github or gitlab? Or setup a gitea/similar website and post it?

I did so with a few of my small stuff.

I just like to release it this way, I didn't have much reason or benefits to releasing it on a git service, hosted or my own

Indeed you are, but if you release it you probably want people to use it, otherwise why release at all? Going github/gitlab way just make that easier.

You are free to do what you want ofc, just my thoughts.

No problem, your thoughts are welcomed

I mainly developed it for myself, and sharing it if others want to do what ever they want to do with it, took many hours to just put together all the readmes and examples and cleaning up the code, I understand everyone wants me to release it on some git platform, but I didn't develop it using any cvs at all and don't plan to do that either