• 2 Posts
  • 116 Comments
Joined 1 year ago
cake
Cake day: June 14th, 2023

help-circle
  • Having had one of the old Windows phones with a keyboard dumped on me at an old workplace, can confirm it’s completely possible for a phone to have a keyboard and be a complete piece of shit.

    A good phone with a good keyboard may have some use cases. If you do a lot of writing but not any more computing power or screen space than a phone has, plus you want to be doing that on the move, then yeah. For me, can shitpost on forums using my phone in my spare time, and dealing with on-call work issues - having multiple tabs of Jira and Slack open, for instance - just isn’t really practical on a small screen.

    If your job is very email-centric, then yeah, sure. Blackberry were very good for just having the stuff you need - email, vpn, ‘corporate’ office documents - in a form that worked.




  • My workplace is a strictly BitBucket shop, was interested in expanding my skillset a little, experiment with different workflows. Was using it as a fancy ‘todo’ list - you can raise tickets in various categories - to remind myself what I was wanting to do next in the game I was writing. It’s a bit easier to compare diffs and things in a browser when you’ve been working on several machines in different libraries than it is in the CLI.

    Short answer: bit of timesaving and nice-to-haves, but nothing that you can’t do with the command line and ssh. But it’s free, so there’s no downside.


  • Ah, nice. Had been experimenting with using my Raspberry Pi 3B as my home Git server for all my personal projects - easy sync between my laptop and desktop, and another backup for the the stuff that I’d been working on.

    Tried running Gitea on it to start with, but it’s a bit too heavy for a device like that. Forgejo runs perfectly, and has almost exactly the same, “very Github inspired” interface. Time to run some updates…



  • Most common example would be a bicycle, I think - your pedals tighten on “in the same direction the wheel turns” as you look at them. So your left pedal has left-hand thread, and goes on and comes off backwards.

    The effect of precession also means that you can tighten the pedals on finger tight and a good long ride will make them absolutely solid - need to bounce up and down on a spanner to loosen them.



  • addie@feddit.ukto196@lemmy.blahaj.zoneI don't make the (rule)s
    link
    fedilink
    arrow-up
    19
    arrow-down
    2
    ·
    1 month ago

    I feel that ‘gender’ is probably a misleading term for the languages that have ‘grammatical gender’, it rarely has anything to do with genitalia. ‘Noun class’, where adjectives have to decline to agree with the class would fit better in most cases.

    English essentially does not have decline adjectives, except for historical outliers like blond/e where no-one much cares if you don’t bother, and uses his / hers / its / erc using a very predictable rule. So no ‘grammatical gender’.





  • We’ve found it to be the “least bad option” for DnD. Have a Discord window open for everyone to video chat in, have a browser window open with Owlbear Rodeo or Foundry / Forge for your tokens and character sheets, all works smoothly enough. The text chat is sufficient for sending the DM a private message; for group chat to share art of the things you’ve just run into or organise the next session.

    Completely agree that for anything “less transient”, then the UX is beyond awful and trying to find anything historical is a massive PITA.



  • Dark Souls’ implementation is something special. Censors your name based on the language settings you have in place at the time, voice-over dialogue remains in English. So change your system language to either another language you know, or play it a few times so you know what things are, and then put the most offensive shit in as your character name you like.


  • addie@feddit.ukto196@lemmy.blahaj.zoneRule
    link
    fedilink
    arrow-up
    13
    ·
    2 months ago

    It’s been a perpetual source of surprise to me that curry houses are so ‘non-specific’. Pakistan and India together make about 1.7 billion people, about a third of the planet’s population, and I’d have thought an easy way to distinguish a restaurant would be to offer something more region-specific, but it’s fairly rare.

    Here in the UK, the majority of curry houses are Bangladeshi - used to be the vast majority, now it’s more like 2/3rds. We’ve a couple of ‘more specific’ chains - both Bundobust and Dishoom do Mumbai-style, and they’re both fantastic - and there’s a few places that do well with the ‘naturally vegan’ cuisines, but mostly you can go in to a restaurant and expect the usual suspects will be on the menu.

    Same goes for Chinese restaurants - I don’t believe that a billion people all eat the same food, it’s too big a place for the same ingredients to be in season all the time. Why are they not more specific, more often?


  • When I was still dual-booting Windows and Linux, I found that “raw disk” mode virtual machines worked wonders. I used VirtualBox, so you’d want a guide somewhat like this: https://superuser.com/questions/495025/use-physical-harddisk-in-virtual-box - other VM solutions are available, which don’t require you to accept an agreement with Oracle.

    Essentially, rather than setting aside a file on disk as your VM’s disk, you can set aside a whole existing disk. That can be a disk that already has Windows installed on it, it doesn’t erase what you have. Then you can start Windows in a VM and let it do its updates - since it can’t see the bootloader from within the VM, it can’t fuck it up. You can run any software that doesn’t have particularly high graphics requirement, too.

    I was also able to just “restart in Windows” if I wanted full performance for a game or something like that, but since Linux has gotten very good indeed at running games, that became less and less necessary until one day I just erased my Windows partition to recover the space.



  • Cheaper for now, since venture capitalist cash is paying to keep those extremely expensive servers running. The AI experiments at my work (automatically generating documentation) have got about an 80% reject rate - sometimes they’re not right, sometimes they’re not even wrong - and it’s not really an improvement on time having to review it all versus just doing the work.

    No doubt there are places where AI makes sense; a lot of those places seem to be in enhancing the output of someone who is already very skilled. So let’s see how “cheaper” works out.


  • PS3 most certainly had a separate GPU - was based on the GeForce 7800GTX. Console GPUs tend to be a little faster than their desktop equivalents, as they share the same memory. Rather than the CPU having to send eg. model updates across a bus to update what the GPU is going to draw in the next frame, it can change the values directly in the GPU memory. And of course, the CPU can read the GPU framebuffer and make tweaks to it - that’s incredibly slow on desktop PCs, but console games can do things like tone mapping whenever they like, and it’s been a big problem for the RPCS3 developers to make that kind of thing run quickly.

    The cell cores are a bit more like the ‘tensor’ cores that you’d get on an AI CPU than a full-blown CPU core. They can’t speak to the RAM directly, just exchange data between themselves - the CPU needs to copy data in and out of them in order to get things in and out, and also to schedule any jobs that must run on them, they can’t do it themselves. They’re also a lot more limited in what they can do than a main CPU core, but they are very very fast at what they can do.

    If you are doing the kind of calculations where you’ve a small amount of data that needs a lot of repetitive maths done on it, they’re ideal. Bitcoin mining or crypto breaking for instance - set them up, let them go, check in on them occasionally. The main CPU acts as an orchestrator, keeping all the cell cores filled up with work to do and processing the end results. But if that’s not what you’re trying to do, then they’re borderline useless, and that’s a problem for the PS3, because most of its processing power is tied up in those cores.

    Some games have a somewhat predictable workload where offloading makes sense. Got some particle effects - some smoke where you need to do some complicated fluid-and-gravity simulations before copying the end result to the GPU? Maybe your main villain has a very dramatic cape that they like to twirl, and you need to run the simulation on that separately from everything else that you’re doing? Problem is, working out what you can and can’t offload is a massive pain in the ass; it requires a lot of developer time to optimise, when really you’d want the design team implementing that kind of thing; and slightly newer GPUs are a lot more programmable and can do the simpler versions of that kind of calculation both faster and much more in parallel.

    The Cell processor turned out to be an evolutionary dead end. The resources needed to work on it (expensive developer time) just didn’t really make sense for a gaming machine. The things that it was better at, are things that it just wasn’t quite good enough at - modern GPUs are Bitcoin monsters, far exceeding what the cell can do, and if you’re really serious about crypto breaking then you probably have your own ASICs. Lots of identical, fast CPU cores are what developers want to work on - it’s much easier to reason about.