Scrath

joined 2 years ago
[–] Scrath@lemmy.dbzer0.com 8 points 5 days ago (3 children)

You can't just say that without providing a backstory dude

[–] Scrath@lemmy.dbzer0.com 3 points 1 week ago

I tried their editor a few times and found it a bit hard to get used to coming from VS Code. The thing that sank the ship for me honestly was when I found out that they don't support PDF previews in the editor. Given that I use VS Code to also write LaTeX documents, that is an essential feature for me that they seemed very opinionated about not including based on the issues I've seen on their git

[–] Scrath@lemmy.dbzer0.com 5 points 1 week ago

I've seen that a couple of times on youtube but always thought it won't look aa convincing in reality.

Also, projectors have really high power usage so I wouldn't be comfortable leaving one running that long effectively being a screensaver

[–] Scrath@lemmy.dbzer0.com 2 points 1 week ago

"Maybe god put it there to help us get started easier on our modern amenities and now that we have better technologies we should move away from destroying gods creation."

[–] Scrath@lemmy.dbzer0.com 1 points 1 week ago

My first internship I kind of landed by talking to a company representative at a career fair organized by my university and then mentionjng said guys name in my online application. Not sure how much impact that actually had but I guess it might have helped.

My second internship I found via linkedin. I just looked for companies there that had somilar positions or actually posted internship positions and then went to their website to apply and look for possible other open positions I could send a slightly altered letter from my first application to. If a company sounded interesting but didn't have any internships on offer I sometimes just sent them an unsolicited application. Honestly I don't remember how much feedback I got on those though.

Should be obvious but I also didn't limit myself to looking for companies close to where I lived at the time. In the end I had to move literally to the other end of the country.

[–] Scrath@lemmy.dbzer0.com 3 points 1 week ago* (last edited 1 week ago)

I didn't read the book but watched the movie instead and I don't know how much is different between the two.

It needed a bit of patience to get through it to be honest. In my opinion, it's more a kind of book/movie you read/watch because of its message, not because it's particularly entertaining.

In a way, I guess that could be seen as part of the message if you look at it as a warning of what could be instead of as a story that is fun to experience.

[–] Scrath@lemmy.dbzer0.com 2 points 3 weeks ago

I'm not that well informed on the specifics of the topic but I would say that AI has a lot of potential to do good in medical applications. I believe there was quite a bit of research into detecting various forms of cancer earlier and more reliably by using neural networks.

[–] Scrath@lemmy.dbzer0.com 1 points 4 weeks ago

A while back I trained a small LSTM based neural net to classify the power phases of a device I work on based on their current consumption over time.

The model worked seemingly great and it took a while for me to notice that it did not catch every phase perfectly.

Yesterday I created a larger and more complex CNN based model on the recommendation of my coworkers which I trained over night since I had to use my work laptop. When applying it to my real data I ran out of RAM. After fixing this issue and getting it to run, it misclassified far too many samples.

I spent the rest of the day building an algorithmic solution that has yet to mislabel a single sample.

This isn't really all that relevant to the post I guess but I found it a nice reminder to myself to actually think about a problem instead of throwing brute force at it and hoping it will solve it. As a side benefit, I can now actually explain why my data is classified the way it is instead of pointing at a black box. There are definitely usecases for AI but you should know enough to recognize when an algorithmic approach is better suited.

[–] Scrath@lemmy.dbzer0.com 5 points 1 month ago

Two rack rails bolted together with a power strip and a tray holding my server mini PC. My router is bolted on as well to act as a switch for everything while also providing Wifi to my phone and laptop

[–] Scrath@lemmy.dbzer0.com 1 points 1 month ago

I kind of railroaded myself into using calibre unfortunately.

I have a very specific filenaming scheme which I originally came up with back when I only used folders for organising my books in order to group together books that belong to a series but where the series is part of a larger universe.

Basically my folder structure is {World}/{Reading Order}; {Series} #{Series_Index} - {Title} - {Author}

On my kobo I have the autoshelf plugin installed which automatically parses this information when I add books and groups them together by world while filling out the series information.

In order to properly make use of this system I need to use Calibre custom columns and be able to export the books I want with this specific name format. I have yet to find a program other than calibre that would support this.

It would probably be smarter for me to reorganize my books at some point but I really like being able to basically drop a ton of books at once onto my reader using SFTP and as far as I can tell all common options rely on manually downloading the books, sending them directly to the reader or pulling them from their internal file storage in whatever form the application stores them...

I do like Audiobookshelf for the ability to add a book to multiple series, but the missing mass export functions stop me from switching

[–] Scrath@lemmy.dbzer0.com 2 points 1 month ago* (last edited 1 month ago)

I name mine after greek and roman gods.

My NAS is bamed Hestia, the goddess of the bearth and home.

My docker server is called Poseidon due to the sea iconography of docker. My second iteration of my docker server where I tried playing around with podman I called Neptune.

I briefly had a Raspberry Pi for experimenting with some stuff which was called Eileithyia, the goddess of childbirth.

My Proxmox machine on which pretty much all ky other servers are run as VMs is called Atlas, as the Titan holding up my personal network.

I also have a truenas VM which I boringly called truenas...

[–] Scrath@lemmy.dbzer0.com 2 points 1 month ago (1 children)

How high is the power bill? I considered getting some more smaller drives but I figured it is more power efficient in the long term to buy bigger HDDs, not to mention that I only have 4 disk slots

 

Hello everyone, I recently tried switching my docker torrent client setup from haugene/transmission-openvpn to linuxserver/qbittorrent with gluetun for my VPN.

I have gluetun set up to use port forwarding with ProtonVPN which assigns a random port on every connection. Gluetun provides a VPN_PORT_FORWARDING_UP_COMMAND which can be used in this scenario to update the port used by qbittorrent. While I had issues with the example command in the gluetun wiki to do this, I eventually managed using a bash script I found in another forum.

My issue now is that my server shuts down for the night to reduce noise and after restarting, even though I have the container startup order set up, qbittorrent is no longer reachable on its webinterface. The logs do not indicate any issue though.

As far as I can tell, the stack as I have set it up is extremely finnicky in terms of startup order and time. If I start gluetun first and then wait too long before starting qbit, the port update script will fail because there is obviously no target for it. If qbit is up and running before gluetun is done, I typically can't access its webinterface for some reason and the network interface used by gluetun will set itself to loopback.

The result of this is that basically every morning I have to start and restart the containers in the stack a couple of times until I can access the interface and ensure that the port and network interface of qbit are configured correctly.

If anyone has a similar setup working that they could share or maybe another solution to my current issue that would be great. Thanks.

This is my docker-compose stack for the new setup

version: '3'
services:
  gluetun:
    image: qmcgaw/gluetun
    container_name: gluetun
    restart: unless-stopped
    cap_add:
      - NET_ADMIN
    ports:
      - 8080:8080 # qbittorrent webinterface
      # - 6881:6881 # qbittorrent, only needed without port-forwarding
    environment:
      - VPN_SERVICE_PROVIDER=protonvpn
      - OPENVPN_USER=${OPENVPN_USERNAME}
      - OPENVPN_PASSWORD=${OPENVPN_PW}
      - SERVER_COUNTRIES=Switzerland
      - VPN_PORT_FORWARDING=on
      - PORT_FORWARD_ONLY=on
      - TZ=Europe/Berlin
      # From gluetun wiki: https://github.com/qdm12/gluetun-wiki/blob/main/setup/advanced/vpn-port-forwarding.md
      - VPN_PORT_FORWARDING_UP_COMMAND=/bin/sh -c 'sh /gluetun/update-port.sh "{{PORTS}}"'
      - VPN_PORT_FORWARDING_DOWN_COMMAND=/bin/sh -c 'echo "Execution port forwarding down command" && wget -O- -nv --retry-connrefused --post-data "json={\"listen_port\":0,\"current_network_interface\":\"lo\"}" http://127.0.0.1:8080/api/v2/app/setPreferences'
      - QBIT_ADDRESS=http://localhost:8080/
      - QBIT_USERNAME=${QBIT_USER}
      - QBIT_PASSWORD=${QBIT_PW}
    volumes:
      - /mnt/truenas/qbittorrent/update_port.sh:/gluetun/update-port.sh
    labels:
      - "com.centurylinklabs.watchtower.enable=true" # Auto update using watchtower

  qbittorrent:
    image: lscr.io/linuxserver/qbittorrent:latest
    container_name: qbittorrent
    environment:
      - PUID=1000
      - PGID=1000
      - TZ=Etc/UTC
      - WEBUI_PORT=8080
      - TORRENTING_PORT=6881
    network_mode: "service:gluetun"
    depends_on:
      - gluetun
        # condition: service_healthy
        # restart: true
    volumes:
      - /home/poseidon/qbittorrent:/config
      - /mnt/truenas/qbittorrent:/downloads
    # ports:
    #   - 8080:8080
    #   - 6881:6881
    #   - 6881:6881/udp
    restart: unless-stopped

And this is the script I use for updating the qbittorrent ports

#!/bin/sh
# update-port.sh
port="$1"
retries="${UPDATE_PORT_RETRIES:-5}"
interval="${UPDATE_PORT_RETRY_INTERVAL:-10}"

echo "Attempting to update qBittorrent port to $port..."

for i in $(seq 1 "$retries"); do
  response=$(wget --quiet --save-cookies=/tmp/cookies.txt --keep-session-cookies \
                  --post-data="username=$QBIT_USERNAME&password=$QBIT_PASSWORD" \
                  --header="Referer: $QBIT_ADDRESS" \
                  "$QBIT_ADDRESS/api/v2/auth/login" -O -)

  if [ "$response" = "Ok." ]; then
    break
  fi

  echo "Login attempt $i/$retries failed. Retrying in $interval seconds..."

  sleep "$interval"
done

set -e

if [ "$response" != "Ok." ]; then
    echo "Unable to log in to qBittorrent."
    rm -f /tmp/cookies.txt

    exit 1
fi

wget --quiet --load-cookies=/tmp/cookies.txt \
     --post-data="json={\"listen_port\": $port, \"current_network_interface\":\"$VPN_INTERFACE\", ,\"random_port\":false,\"upnp\":false}" \
     "$QBIT_ADDRESS/api/v2/app/setPreferences" -O /dev/null

rm -f /tmp/cookies.txt

echo "qBittorrent port updated successfully to $port."

Update

After updating my stack based on the recommendation from hyphen612 it has been running flawlessly for a few days. This is my new docker-compose file. The extra script I used before for updating the port has been retired.

version: '3'
services:
  gluetun:
    image: qmcgaw/gluetun
    container_name: gluetun
    restart: unless-stopped
    cap_add:
      - NET_ADMIN
    ports:
      - 8080:8080 # qbittorrent
      # - 6881:6881 # deluge or qbittorrent, only needed without port-forwarding
    environment:
      - VPN_SERVICE_PROVIDER=protonvpn
      - OPENVPN_USER=${OPENVPN_USERNAME}
      - OPENVPN_PASSWORD=${OPENVPN_PW}
      - SERVER_COUNTRIES=Switzerland
      - VPN_PORT_FORWARDING=on
      - PORT_FORWARD_ONLY=on
      - IPV6=off
      - TZ=Europe/Berlin
      - VPN_PORT_FORWARDING_UP_COMMAND=/bin/sh -c 'wget -O- --retry-connrefused --post-data "json={\"listen_port\":{{PORTS}}}" http://127.0.0.1:8080/api/v2/app/setPreferences 2>&1'
      - HEALTH_SUCCESS_WAIT_DURATION=20s
      - HEALTH_TARGET_ADDRESSES='1.1.1.1:443'
    healthcheck:
      test: ["CMD", "/gluetun-entrypoint", "healthcheck"]
      interval: 10s
      timeout: 10s
      start_period: 20s
      retries: 10 #has internal fix mechanism
    labels:
      - "com.centurylinklabs.watchtower.enable=true" # Auto update using watchtower

  qbittorrent:
    image: lscr.io/linuxserver/qbittorrent:latest
    container_name: qbittorrent
    environment:
      - PUID=1000
      - PGID=1000
      - TZ=Etc/UTC
      - WEBUI_PORT=8080
      - TORRENTING_PORT=6881
    network_mode: "service:gluetun"
    depends_on:
      gluetun:
        condition: service_healthy
        # restart: true
    volumes:
      - /home/poseidon/qbittorrent:/config
      - /mnt/truenas/qbittorrent:/downloads
    restart: unless-stopped

Update 2

The issue with qbittorrent not being reachable despite the container being healthy did not go away with the previous update. What finally fixed the issue for me was changing the tag of the container image. I am now using the lscr.io/linuxserver/qbittorrent:libtorrentv1 image instead and that seems to work so far.

 

Hello everyone, I am currently looking for a software solution to use my home server as a DLNA renderer which can output audio to my stereo amplifier.

The only solution I found was called gmrender-resurrect which seems like it would do exactly what I want but I was unable to get a docker container of it working. While I was able to find and connect to the DLNA Renderer, playback would fail every time and I was unable to get any information from the logs regarding why.

Do any of you know another solution to stream audio from my phone to my server (I am using Symfonium on the phone side)? Ideally it would be something I can deploy as a docker container on my server.

Thanks.

 

Hello everyone, I am currently trying to set up a kmonad config file to replace the autohotkey script I used on windows. My goal is simply to use the right alt key in combination with a,o,u and so on to type german umlaut characters like ä,ö,ü, etc.

So far I am having trouble even getting kmonad to run the config. I guess I probably misunderstand how this is supposed to work significantly. My initial config file was generated by ChatGPT since I had no idea where to even start.

This is my current config file

(defcfg
  input  (device-file "/dev/input/by-path/platform-i8042-serio-0-event-kbd")
  output (uinput-sink "kmonad_keyboard")
  fallthrough true
  allow-cmd true
)

(defsrc
  ralt a o u s lsft
)

(deflayer german
  ralt-a "ä"
  ralt-o "ö"
  ralt-u "ü"
  ralt-s "ß"
  ralt-shift-a "Ä"
  ralt-shift-o "Ö"
  ralt-shift-u "Ü"
)

Any help would be appreciated.

44
submitted 1 year ago* (last edited 1 year ago) by Scrath@lemmy.dbzer0.com to c/electronics@discuss.tchncs.de
 

Hello everyone, I recently built a small distribution board to distribute 5V to multiple components for use in a robotics project. I made each output switchable with an individual switch and an LED to indicate the current state. When I went to test it using a lab power supply I noticed that the LEDs would start flickering weirdly when I turned them off and on again.

https://imgur.com/a/zaSCUby

As it turns out, the LEDs, which I found in my dads old parts in a bag labeled TLBO 5410, are apparently blinking LEDs. I found a datasheet for TLBR5410 LEDs which seem pretty much identical to what I have accidentally used.

Apparently these LEDs are made to operate directly from a 5V supply without an additional current limiting resistor (it is already built in) and are made to continuously blink at a frequency of 3Hz.

Because I thought I was using standard LEDs I added a series resistor causing them to behave weirdly due to low voltage. For comparison, this is how they are supposed to act: https://imgur.com/a/fXlcEDs

 

Hello everyone, I have another question regarding reverse-proxying again, specifically for the linuxserver.io jellyfin image.

On the dockerhub page for this image there are 4 ports listed which should be exposed:

  • 8096 for the HTTP Web UI
  • 8920 for the HTTPS Web UI
  • 7359/udp for autodiscovery of jellyfin from clients
  • 1900/udp for service discovery from DLNA and clients

Additionally there is also an environment variable JELLYFIN_PublishedServerUrl which is for "Setting the autodiscovery response domain or IP address". I currently have that set to my subdomain https://jellyfin.mydomain.com though I am not sure if that is correct.

I already have a reverse-proxy set up allowing me to access my servers webinterface under https://jellyfin.mydomain.com without exposing the https port on the container. What I am unsure about now however, is what to do with the two ports for UDP traffic.

By my understanding, a reverse-proxy will only forward traffic which comes to the ports 80 for http and 443 for https. Those are also the only ports my reverse-proxy container has exposed alongside the management interface. As such the 2 udp ports will not be reachable under my jellyfin domain.

How can I change this or is this even an issue?

10
submitted 2 years ago* (last edited 2 years ago) by Scrath@lemmy.dbzer0.com to c/selfhosted@lemmy.world
 

Hello, I have a question regarding the usage of a reverse-proxy which is part of a docker network.

I currently use Nginx Proxy Manager as a reverse-proxy for all my services hosted in docker. This works great since I can simply forward using each containers name. I have some services however (e.g. homeassistant) which are hosted separately in a VM or using docker on another device.

Is it possible to use the same reverse-proxy for those services as well? I haven't found a way to forward to hosts outside of the proxies docker network (except for using the host network setting which I would like to avoid)

view more: next ›