• 0 Posts
  • 7 Comments
Joined 1 year ago
cake
Cake day: June 10th, 2023

help-circle
  • I tend to agree, but there are two issues working agaonst Star Trek.

    1. Successful media appeals to broad audiences by having something to appeal to every demographic. (E.g. Don’t like politics? Stay for the lasers.)
    2. Good sci-fi (arguably stories in general) gives the best representation of both sides of a conflict, and lets them compete on their merits. So it’s possible to resonate with one side, then miss the critique (e.g. due to modest writing or selective hearing).

    So while Star Trek tends to show progressive values winning in the end, many people can enjoy other aspects (e.g. military stories, relationships, and action) while ignoring the upshot.






  • For LLMs, I’ve had really good results running Llama 3 in the Open Web UI docker container on a Nvidia Titan X (12GB VRAM).

    For image generation tho, I agree more VRAM is better, but the algorithms still struggle with large image dimensions, ao you wind up needing to start small and iterarively upscale, which afaik works ok on weaker GPUs, but will gake problems. (I’ve been using the Automatic 1111 mode of the Stable Diffusion Web UI docker project.)

    I’m on thumbs so I don’t have the links to the git repos atm, but you basically clone them and run the docker compose files. The readmes are pretty good!