Next evolution, just a one line bash script.

  • linearchaos@lemmy.world
    link
    fedilink
    English
    arrow-up
    127
    ·
    10 months ago

    Me: install it, doesn’t work, read the docs, screw with all the missing things, doesn’t work, read the forms, install something else I missed, doesn’t work, find more forums, find the right answer, patch it up, get it working, figure out that the application is slow, missing critical features, and really just doesn’t do what I needed to do.

    • deweydecibel@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      ·
      10 months ago

      Or it does work, and then I never actually end up using it again.

      And then months later I’ll have to do something similar and I’ve forgotten I even installed something that can do that, so I install another related thing.

    • Punkie@lemmy.world
      link
      fedilink
      arrow-up
      6
      ·
      10 months ago

      really just doesn’t do what I needed to do.

      This has been my experience, or sort of does what I want it to do, but I have to rethink what I need it to do instead of something really simple. Like a “new type of shared file system” that replaces NFS/Windows sharing. So instead of files in a standard file system one can manage with a file browser, it has “indexed” your files in such a way that the actual files are renamed into data chunks, and one “finds” files by their non-intuitive search engine that can’t do even basic search engine tricks like “AND/OR” searches, wildcards, and the results are hit and miss. “But it’s faster and more elegant!” So how do you restore from backup when the system fails? “When the system does whatnow?”

      Yeah, no thanks. I can recover files from a file system much easier than some proprietary encoded bullshit fronted with a bad search engine over a proprietary and buggy index.

      • linearchaos@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        10 months ago

        I asked the other week if anyone made a system that left files alone and just indexed them and gave you a place to store meta without moving them. Options do seem to exist, but they need LOTS of extra work

    • JasonDJ@lemmy.zip
      link
      fedilink
      arrow-up
      2
      ·
      10 months ago

      I have a couple scripts running in containers. I’m too lazy to do a proper CI build and I only run it on one system anyway so I just got pull and docker build. Sorry if i offend you.

      Though if i keep this up I’ll probably just start to use my generic python container and commit a shell script to docker run with . mounted to /project or something, and set the entry point to pip install project/requirements.txt; python project/main.py.

  • Tja@programming.dev
    link
    fedilink
    arrow-up
    38
    ·
    10 months ago

    I was blown away how a relatively unknown project like immich provides one Docker compose file to bring up a whole self-hosted Google photos photo management suite, complete with tagging, mapping, transcoding and semantic search. Local, offline semantic search.

        • 1337A
          link
          fedilink
          arrow-up
          10
          ·
          10 months ago

          I’m having a hard time getting into software that doesn’t use traditional folder structures. I wish I could point immich to my NAS Pictures folder and have multiple places to seen the same exact files, immich when I want, nextcloud maybe, or just a file browser pointed to my NAS.

          I’m having the same issue with paperless-ngx. I set it up and it’s cool, but why can’t I just point it to my Documents folder? I’m getting all of my 2023 taxes ready and now I have to upload them into two places, paperless-ngx for my records and Nextcloud so I can ultimately share a link with my accountant.

          Am I old man yelling at cloud-ing right now? Why get rid of basic folder structures I just don’t get it

          • Tja@programming.dev
            link
            fedilink
            arrow-up
            4
            ·
            10 months ago

            Immich does use traditional folders (it’s one of the parameters) but it does a copy there. It stores the pictures in multiple formats/resolutions and does some caching and whatnot, so it probably wouldn’t be a nice sight. Right now. On the other hand it does organize your collection in the pattern that you give it, for instance year/month/day_hour_minute.jpg. It could probably be implemented in the future to have the meta-images somewhere hidden and leave the collection alone, but that all requires effort and coordination. And everyone has their photos organized in their own little snowflake method.

            Having said that, I totally understand your point. I’ll yell at the cloud with you.

          • rambos@lemm.ee
            link
            fedilink
            arrow-up
            4
            ·
            10 months ago

            There are external libraries in immich. You point folder path to immich and you can see everything, but you cant eddit or delete

        • Samsy@lemmy.mlOP
          link
          fedilink
          arrow-up
          1
          ·
          10 months ago

          I do the same, but I just upload them to paperless, there is a cronjob doing the paperless export and rclone new exports to my nextcloud.

    • dutchkimble@lemy.lol
      link
      fedilink
      arrow-up
      6
      ·
      10 months ago

      I got into the idea of selfhosting my photos, and wanted facial recognition to search them. I also wanted a selfhosted chat server. Nextcloud came as the obvious choice that did both. After days of tinkering and fixing errors in the log one by one, everything worked except the facial recognition which said everything was fine but the faces just didn’t get recognised. Also the mobile experience for Memories wasn’t the best. Then luckily I came across immich and it was up and running in about 5-6 mins of configuration max, and it has better facial recognition than the main big commercial option. Insane. For chat I got a synapse server with coturn which also took about 15 mins of docker composing and setting configurations/accounts to my liking.

      (I still think Nextcloud is cool but it’s overkill and loaded with too many features I don’t need + installation is a task & documentation/online support communities are scattered between the various methods of installation)

  • Juki@lemmy.world
    link
    fedilink
    arrow-up
    40
    arrow-down
    2
    ·
    10 months ago

    I’m the opposite because I’ve had nothing but bad luck with docker. I should really spend more time with it but ugh

    • CosmicTurtle@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      1
      ·
      10 months ago

      It’s definitely worth learning. I had the damnedest time with docker until I went to a meetup and had someone ELI5 to me. And it wasn’t that I wasn’t technical. I just couldn’t wrap my head around so many layers of extraction.

      The guy was very patient with me and helped me get started with docker compose and the rest is history.

    • pHr34kY@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      10 months ago

      I’m like that. It feels like a total waste of resources, and introduces unneeded complexity for backup, updates, file access, networking and general maintenance.

      I would take a deb repo over docker any day of the week.

  • rushaction@programming.dev
    link
    fedilink
    arrow-up
    36
    arrow-down
    1
    ·
    10 months ago

    For me it’s more like new interesting self hosted project and then find out it’s only distributed as a docker container without any proper packaging. As someone who runs FreeBSD, this is a frustration I’ve run into with quite a number of projects.

    • zaphod@lemmy.ca
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      1
      ·
      edit-2
      10 months ago

      Eh even as a Linux admin, I prefer hand installs I understand over mysterious docker black boxes that ship god knows what.

      Sure, if I’m trialing something to see if it’s worth my time, I’ll spin up a container. But once it’s time to actually deploy it, I do it by hand.

        • JasonDJ@lemmy.zip
          link
          fedilink
          arrow-up
          5
          arrow-down
          1
          ·
          edit-2
          10 months ago

          Sorry but IMO that’s FUD.

          The reliance on it legitimately prevents the issues that it’s likely to cause. It’s made to be both idempotent and ephemeral.

          Give an example of a Python project. You make a venv and do all your work in there. You then generate a requirements with all the versions pinned. You start build a container on a pinned version of alpine, or Ubuntu, or w/e. Wherever possible, you are pinning versions.

          With best practices applied, result is that the image will be functionally the same regardless of what system builds it, though normally it gets built and stored on a registry like Docker Hub.

          The only chance a user has to screw things up is in setting environment variables. That’s no different than ever before. At least now, they don’t have to worry about system or language-level dependencies introducing a breaking change.

      • JasonDJ@lemmy.zip
        link
        fedilink
        arrow-up
        4
        arrow-down
        1
        ·
        10 months ago

        If it’s an open-source project, usually the dockerfiles are available for reading.

        Do you audit every line of code that you run in production? If you are trying some new python/django/sql app, are you reviewing all that?

        I’d assume with a python based project, you’d be able to at least look at requirements and tell there’s something that sets off red flags. And you are either familiar/trust the maintainer, or you are reviewing the actual python itself?

        Beyond that, the dockerfile is essentially just installation instructions for getting it running on a virgin system of X distribution. I wouldn’t call that a black box.

        If the container isn’t part of an open source project, then this is a moot point then. The project itself is a black box.

        • zaphod@lemmy.ca
          link
          fedilink
          English
          arrow-up
          5
          ·
          edit-2
          10 months ago

          You do you. Speaking for myself, I prefer to understand and be able to trivially inspect and modify the moving parts in the things I deploy so I have a snowball’s chance in hell of debugging and fixing things when something inevitably goes wrong.

            • zaphod@lemmy.ca
              link
              fedilink
              English
              arrow-up
              2
              ·
              10 months ago

              And all I see is someone taking this conversation way too personally.

              • JasonDJ@lemmy.zip
                link
                fedilink
                arrow-up
                1
                arrow-down
                3
                ·
                10 months ago

                You sound like someone who doesn’t want to save 10 minutes of work every day because it might cost you half an hour every month.

      • rushaction@programming.dev
        link
        fedilink
        arrow-up
        7
        ·
        10 months ago

        Virtualization in general? Sure, I can. I’ve tried it a bit with bhyve. But it’s definitely a lot heavier since I’m now running a full Linux os and dedicating resources to it to run docker just to run a python or node app.

        Learning the project is in Go though is a sigh of relief. Professionally I’ve moved to Go (from Python) just because it’s so damn easy to build and distribute.

        I just wish there was better support for the other *nix’s. While the language support them just fine, docker on the other hand strangles it. =(

  • Harvey656@lemmy.world
    link
    fedilink
    arrow-up
    33
    ·
    10 months ago

    This oh my God. Just the other day I tried to install a project off git, it had a nice little .bat file to install all the requirements except half if them just didn’t exist or were so niche I couldn’t find anything on them after searching. Would love more dockers please.

      • Samsy@lemmy.mlOP
        link
        fedilink
        arrow-up
        22
        ·
        10 months ago

        Ah well, maybe he used to much aliases? starts sweating in penguin costume

      • Harvey656@lemmy.world
        link
        fedilink
        arrow-up
        7
        ·
        10 months ago

        Naw they only had windows projects. I run all my stuff through VMware. Gotta have windows for stupid easy anti-cheat. Trust me I only use it when I have to, please put the gun down mr railcar!

        • lengau@midwest.social
          link
          fedilink
          arrow-up
          4
          ·
          10 months ago

          Cursed project: provides .ps1 file for *nix configuration, .sh for Windows configuration (using git bash)

  • Evilschnuff@feddit.de
    link
    fedilink
    arrow-up
    28
    ·
    10 months ago

    I believe this to be true for nearly all products. It has to be super simple to test, because you need to assess if it fits your needs. The mental model for a priori assessment is not strong enough usually.

  • Crow@lemmy.world
    link
    fedilink
    English
    arrow-up
    27
    ·
    10 months ago

    It’s because I’ve seen What people can do with a simple docker container that I completely agree. It’s too nice to go back.

      • brophy@lemmy.world
        link
        fedilink
        arrow-up
        8
        arrow-down
        1
        ·
        10 months ago

        There’s rootless docker, or podman, or numerous other container runtimes. The beauty in containers is separating concerns. How you choose to run it, root or rootless, is up to you in all but the nichest of scenarios.

        • lemmyingly@lemm.ee
          link
          fedilink
          arrow-up
          3
          ·
          10 months ago

          Rootful docker and rootless docker can be run at the same time on the same machine too. So projects that require root privileges can still work on a machine where most other projects run as rootless.

  • Possibly linux@lemmy.zip
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    10 months ago

    Building a docker container isn’t normally to hard. I usually will create a PR with a dockerfile and docker compose

  • TeaEarlGrayHot@lemmy.ca
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    10 months ago

    I was so surprised by how easy it was to install Yacy–I’d thought a self-hosted search engine would be tough, but I made a docker-compose file and pointed my reverse proxy to the server, works perfectly so far!

  • mlg@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    6
    ·
    10 months ago

    Yeah no thanks I actually enjoy customizing my installs and not relying on docker for config management which it really shouldn’t be used for.

    Only container I have that was well worth it is the OSX vm which makes it easy to swap versions and options without having to coax the crappy apple software.

    Which I also only have because I thought it’d be funny to demo bluebubbles to my friends.