How to add webdav storage with http only in web edition?

it auto add as https.

Since it’s unencrypted I assume it’s local, and in this case, you can use any other protocol that works well in lan.

WebDAV should really be a last resort thing.

What server are you connecting to?

macos with wireguard

So, why WebDAV, let alone unencrypted? macOS supports SFTP out of the box, NFS, SMB, etc… Use any of those protocols instead. I would recommend SFTP, as you likely already use SSH, so no extra configuration will be required.

for webdav, i can only bind the service with wireguard’s interface to ensure anyone else can’t visit.
I’m not use ssh by the way.

WebDav is way, way, less secure than SFTP. So if you are planning to expose it this to the internet – don’t.

I don’t know what’s your setup to provide specific recommendations, but if you describe your network (i.e. how is the client connecting to the internet, where is the server – firewalls, any other restrictions, how are you using WireGuard: hosting a server at your server, or using it as a bridge on a VPS and route packets with masquerading, etc., we could suggest something.

OP is using Wireguard for security, WebDav security is irrelevant here. In fact, running SFTP within Wireguard tunnel is a massive overkill - double wrapping only makes things slower without gaining anything.

We don’t know that. I asked for clarifications. It is not obvious that wireguard is end to end in this scenario. It can be used to connect to jump box and expose port there; for example as a workaround if a server is behind restrictive firewall or a provider NAT

You gain performance, stability, and reliability by avoiding WebDAV.

I use WebDAV over wireguard. If your setup’s CPU-bound (e.g. no hardware AES) performance beats SFTP.

Duplicacy works with http WebDAV but you have to edit your settings file directly - the web interface wont accept it.

So stop duplicacy, open duplicacy.json and edit the storage entry to look something like this:

        {
            "name": "Remote",
            "url": "webdav-http://duplicacy@odroid-remote/duplicacy/",
            "encrypted": false,
            "ras_encrypted": false,
            "erasure_coding": "",
            "credentials": {
                "webdav_password": "XXXXXXXXXXXXXXXXXXXXXXX"
            }
        }

But first I’d verify the WebDAV server is working correctly. On linux you can use litmus to test - it should pass all tests except propfind_invalid2, which you can ignore.

If there is misconfiguration that allows someone unauthorized to gain access to another side of encrypted Wireguard tunnel, there could be misconfiguration that allows to gain access to the other end of encrypted SFTP tunnel just as well. This has very little to do with the original OP question.

SFTP can be slow. There is a reason why rclone allows to serve both SFTP and WebDAV, e.g. people who do streaming from rclone for instance tend to use WebDAV.

@sl474118737: I don’t know if you can do it from the GUI, but you can edit :d: preferences file manually. Ideally you’d add some https WebDAV storage (some other server), then edit “storage” key in preferences file from “webdav://username@httpsserver:port” to “webdav-http://username@httpserver:port”.

That’s not at all what I said.

sftp is rarely a bottleneck over the internet. And if it is, and wireguard is indeed used end to end — there are other protocols, like NFS, available. There is no place for WebDAV in this usecase, and in my opinion duplicacy shall drop it altogether: it’s never suitable.

That’s not the reason that is applicable here, and using webdav to serve bulk storage is contrary to the purpose and protocol design. Rclone provides webdav adapter for webdav applications. Streaming, that benefits from http range requests, is one of them. But this is entirely different usecase. I’d say, completely opposite: few large files with range requests vs many small files retrieved as a whole.

It has its uses. serviving a backup target (thousands of small files) is not one of them. This is an overgeneralization.

I am so glad that you’re not the one making these calls. I am using all the things that in your opinion should be dropped from :d: - GDrive, OneDrive, WebDav etc. Works quite well for me, but hey, none of these are ever suitable. /s

1 Like

We had this conversation before. I don’t accept half-measure and “ wrong, but works, quite well” solutions. Only accept well motivated designs, that always work, scale well, and not just in specific circumstances, for specific users, for indeterminate periods of time, with poor alignment of incentives. Doing otherwise provides poor user experience.

Even though a forum is not a reflection of real world usage, you can count topics about issues with e.g. S3 vs e.g. gdrive, to gen some vague idea.

And besides, it did not work anywhere near “quite well” for one user — me. Albeit it did start promising, on a small-ish datasets in simple use-cases. I’ve suffered through it for two years, far too long in the hindsight. Those issues are not fixable, for the reasons discussed before. I have never heard of anyone who had issues with AWS S3 or GCS.

Yes, anything can be made to work, even “quite well”, but I refuse to accept this copout.

Let’s be honest, the only reason you use and tolerate high latency google drive is because it’s “free” hot storage and duplicacy does not support archival tiers. I seriously doubt that it was even a contender had price not been a factor or if archival storage was supported. That, or you have a tiny dataset, as these solitons don’t scale.

So, instead of defending misguided compromises how about advocating for the appropriate for the job scalable design and making the correct choices for the users.

Because the backup program that supports webdav among main protocols screams “I don’t know what i’m doing” or “I don’t care about user experience” or “my users shall be able to shoot themselves in the foot in the worst way possible”. The latter bit is important. Those solitons work somewhat in the beginning but then invariably fall apart, but you are too deep in.

So yes, if duplicacy dropped webdav and google drive tomorrow, some people, including you, will be pissed, some will move to another solution, and some will trust the developer and play along, migrate to the appropriate for the job technology and be better off in the long run.

But if short term profit is the driving force — yes, we will have these posts from users about webdav, google drive, latency, and other, completely avoidable, hurdles.

I’m surprised I have to elaborate on this.

But just look at the modern software landscape, good software is far and between. Why? Because people are content with “quite well”.

Yes, this is an excellent reason to eliminate this as an option for everybody else /s. You know that yours is not the only use case out there, right? By the way, one of my datasets is about 20TB, how much have you tested with?

I don’t even know if I can add anything to that. If price is not a factor, literally every single thing in the world would be different. But price is a factor, in everything. At least for people in the real world.

I am advocating for having choices for users, and some of them might not be just like you (gasp!) and have different use cases and considerations as to what is important or not for them.

1 Like

Under 3TB. Above two it was started getting unbearable. Prune would take weeks and restore version enumeration tens of minutes. Please don’t tell me I had to have fewer versions: deleting data to make slow software happy is not a solution. On the contrary, it is an illustration that this backend does not work.

The problem here is that duplicacy forcing users to hot storage makes them do these compromises. Archival storage is cheap, and is best suited for backup.

And besides price there is “value”. Optimizing price is foolish: lowest cost solutions are notorious at offering poor value.

That’s where we disagree the most. I think users shall have no say in inner workings of the solution. Even if having choices is great on paper, supporting all those choices is often impossible. If I know nothing about backup, and I pay money for a piece of software, i expect it to guide me to the “correct” solution. I don’t want it to force me to neither make mundane choices nor those in the domains I’m not an expert on, nor “learn on my own mistakes”.

Backup software developer is in much better position to decide which storage providers are suitable to provide the designed user experience.

Ideally, there shall be just one backend, (one of the big three), with all the optimizations made to take full advantage of that one backend specifics. Supporting 17 hundreds backends dilutes the quality of each, and increases support non-linearly — crappy backends will generate more support volume.

The software design is already fixed. Some storage provides will objectively work better than others. Hence, there can be one provider chosen that provides best experience. There is no room for user choice other than for the sake of having one. If only one provider was supported we would not be having these chats.

The only reason we wouldn’t be having these chats is that many current users (myself included) would be using some other software. I am done with this conversation.

1 Like

Above I said that some users like yourself, who evidently prefer penny pinching to quality, may end up leaving for the “competitors”. And that’s ok.

You are vastly overestimating number of such users. Had this been my software — I would happily fire all my cost-{preoccupied,driven} users. These users generate least profit, but require tons of support (both directly and indirectly, in the form of development and testing spent on workarounds for bad backends). Been there, done that.

Regardless of the nature of the product, with the fixed amount of available resources you can either make a quality product that is not cheap, or half-hearted half-baked contraption riddled with “user choices”, at bottom of the barrel prices. Not both. You are advocating for the latter. I’m for the former.

I’m still not sure why you think that shifting the burden of making critical domain-specific choices from the vendor to the customer is an acceptable, let alone desirable, thing.

I can imagine SFTP outperforming WebDAV but what’s the concern with stability/reliability?

I’ve run 2 WebDAV storages for about 2 years (with check -chunks as part of the schedule) and the only issue was one corrupted chunk which I suspect was caused by a power outage.

WebDAV protocol design goals (document exchange) are drastically different from what software like duplicacy requires (fast access to a massive arrays of small files), and as a result, there are bottlenecks in “weird” places, slowing down access in the best case, or losing data in the extreme, where e.g. server runs out of ram, drops connection, or writes corrupted data due to some bugs: there is the whole web server driving this under the hood, as opposed to purpose-build engine like sftp.

Here here.

Regardless of whether or not you think users should be given “choice” by any given software developer, this statement alone demonstrates why they absolutely should be given choices at all costs. You (in particular) are not qualified, nor have the experience of others, to judge what is a “wrong” solution.

Case in point - you claim GCD is inadequate for Duplicacy. My vast and flawless experience with it says you’re 100% wrong.

How do we resolve this conflict?

Well how about offer users some bloody “choice” and stop treating everyone like children?

Let’s be honest, the only reason you keep dissing on Google as a provider is you quite clearly have a bias against almost everything they do. They compete with Apple and it pisses you off their customers aren’t as bled dry as you are.

Compromises, such as not verifying backups due of exorbitant egress costs, perhaps?

Advising people to switch to sftp from WebDAV is one thing. (I wonder, how you would have come to your conclusion, had you not the opportunity to test it for yourself.)

Advocating for the removal of all but the protocols and features you personally approve of, is the height of arrogance. (I wonder, how you’d feel if I said ‘archival tier’ storage should never, ever, get added to Duplicacy. And yet, even though I would never use it, I’d be very happy if it did get added.)

And yet despite all this, somehow, this is a wrong solution. Shame on gchen for steering you down this forbidden path :wink: