• 2 Posts
  • 33 Comments
Joined 3 years ago
cake
Cake day: June 15th, 2023

help-circle

  • As someone who bought Half Life 2 when it was released …

    I only remember people being excited about Steam, Web stores weren’t a thing back then and they were the future! (It was the following years of audio and ebook stores locking stuff down and evapourating that taught us to hate it).

    Game/Audio CD DRM hacking the kernel and breaking/massively slowing down your PC was pretty common back then and Steam’ s DRM didn’t do that.

    The HL2 disc installer didn’t require you to install Steam, once installed it asked you to setup Steam and there was a sticker under the DVD with the Steam code for you to enter.

    You were then rewarded with a copy of HL2 Deathmatch and Counterstrike Source.

    Steam wasn’t always on DRM, back then ADSL/DSL was relatively new and alot of people were still stuck on Dial Up modems.

    Steam let you sign in and authorize your games for 30 days at which point you would need to log into Steam again. This was incredibly helpful feature for young me.


  • Basically Epic like every other publisher has created their own launcher/store.

    They aren’t trying to compete on features and instead using profits from their franchise to buy market share (e.g. buying store exclusives).

    The tone and strategy often comes off as aggressive and hostile.

    For example Valve was concerned Microsoft were going to leverage their store to kill Steam. Valve has invested alot in adding windows operability to Linux and ensuring Linux is a good gaming platform. To them this is the hedge against agressive Microsoft business practices.

    The Epic CEO thinks Windows is the only operating system and actively prevents Linux support and revoked Linux support from properties they bought.

    As a linux user, Valve will keep getting my money and I literally can’t give it to Epic because they don’t want it.




  • Nvidia drivers don’t tend to be as performant under linux.

    With AMD instead of using the AMD VLK driver, you would use the RADV (developed largely by valve). Which petforms better.

    Every AMD card under linux supports OpenCL (the driver is more based on graphics card architecture) and you install it very easily. Googling it with windows found pages of errors and missing support.

    Blender supports OpenCL. I bet the 2x improvement is Blender being able to ofload rendering to the AMD graphics card.

    Also this represents the biggest headache in Linux, lots of gamers insist they can only use Nvidia cards. Nvidia treats linux as an afterthought as best or deliberately sabotages things at worse.

    AMD embraced open source and so Linux land is much nicer on AMD (and to a less extent Intel).

    The results here will probably be a DxVK quirk, lots of “Nvidia optimised” games have game engines doing weird things and the Nvidia driver compensates. DxVK has been identifying that to produce “good” vulkan calls.


  • Mint was a reaction to Gnome 3, the unique workflow upset a lot of people and the people behind Mint decided to build Cinnamon desktop (its Gnome 3 made to look/work like Gnome 2). They needed a distribution to build/test their work and so based a distribution off of Ubuntu and called it Mint.

    As a bit of explanation, there are only a few projects which attempt to build an entire linux distribution from scratch. This involves finding code from thousands of sources, work out packaging, etc… We call these ‘base’ distributions, Debian is the base distribution for Ubuntu, Ubuntu is the base distribution for Mint.

    Ubuntu tends to be slightly ahead of Debian in the software versions it uses and automatically enables the ‘non-free’ repositories. Ubuntu tends to push some Canonical specific things like Snaps (which everyone hates)

    I believe Mint rolls the Canonical specific things out of Ubuntu and you get the latest version of Cinnamon.

    Its all a bit…


  • If its for work I would suggest picking a “stable” distribution like Debian, Kubuntu or OpenSuse.

    A lot of people recommend Arch or Fedora but the focus of those is getting the very latest releases, which increases your chance of stuff breaking.

    A lot of people will suggest niche distributions, those can be great for specific needs but generally you will always find Debian/Ubuntu/RHEL support for commercial apps.

    I would also suggest looking at the KDE Desktop, many distributions default to Gnome but it is unique in how it works, KDE (or XFCE) will provide a desktop similar to Windows 11.

    Lastly I would suggest looking at Crossover Linux by Codeweavers.

    Linux has something called WINE, its an attempt to implement the Windows 95 - 11 API’s so windows applications can run on linux.

    WINE is how the Steam Deck/Linux is able to play Windows games. Valve embedded it into Steam and called it “Proton”.

    WINE is primarily developed by Codeweavers and they provide the Crossover application that makes setting up and running a Windows application really easy.

    People will mention Lutris but that has a far higher learning curve.

    There is an application database so you can see in advance if your applications would work: https://appdb.winehq.org/


  • This advice isn’t grounded in reality.

    Management normally defines ways to track and judge itself, these are typically called Key Performance Indicators.

    KPI’s are normally things like contract value growth, new contracts signed, profit margin, etc…

    So if the project manager is meeting or exceeding their KPI’s and you walk up to their boss telling them the PM is failing as basic job functions, the boss won’t care.

    This is because the boss might have set the KPI’s or the boss might also be judged on them. In either situation its to the bosses advantage to ignore you.

    The boss will only care if there is a KPI you can demonstrate the PM failing to meet.

    Every person/group will have various incentives and motivations. To affect change you have to understand what they are.



  • A project manager has responsibility for delivery of a project but they typically lack domain specific knowledge. As a result they can’t directly deliver something, merely ask subject matter experts for advice and facilitate a team to deliver.

    Most PM’s cope with the stress of this position poorly.

    This cartoon is an example of micro management (a common coping mechanisim), the manager has involved themselves in the low level decisions because that gives a sense of control. If a technical team then tell them its a bad decison the team are effectively attacking their coping mechanisim.

    The solution isn’t to tell them their technical idea is terrible, when you’ve fallen down this rabbit hole you have to treat the PM as a stakeholder. They are someone you have to manage, so a common solution is to give them confidence there is a path to delivery, a way to track and understand it.


  • SpaceX are launching 26-52 satellites at a time and have sustained 3 launches a week for most of the year.

    The satellites are in a Low Earth Orbit, without constant thrust, atmospheric drag will force them to re enter earths atmosphere within a few months. This means they aren’t adding to junk in space.

    Unlike Nasa, ULA, Arriannespace, RoscosMos, etc… SpaceX have always performed 2nd Stage Deorbit burns, so they aren’t adding to Space junk by launching either.

    The Low Earth Orbit is to ensure low latency and the need for constant thrust means the satellites have a short life expectancy by design. That is why SpaceX fought to keep the satellites as cheap as possible (e.g. $250k)

    First stage booster reuse and fairing reuse means the majority of the launch cost is the second stage ($15 million).

    The whole lot is privately funded



  • During the pandemic I had some unoccupied python graduates I wanted to teach data engineering to.

    Initially I had them implement REST wrappers around Apache OpenNLP and SpaCy and then compare the results of random data sets (project Gutenberg, sharepoint, etc…).

    I ended up stealing a grad data scientist because we couldn’t find a difference (while there was a difference in confidence, the actual matches were identical).

    SpaCy required 1vCPU and 12GiB of RAM to produce the same result as OpenNLP that was running on 0.5 vCPU and 4.5 GiB of RAM.

    2 grads were assigned a Spring Boot/Camel/OpenNLP stack and 2 a Spacy/Flask application. It took both groups 4 weeks to get a working result.

    The team slowly acquired lockdown staff so I introduced Minio/RabbitMQ/Nifi/Hadoop/Express/React and then different file types (not raw UTF-8, but what about doc, pdf, etc…) for NLP pipelines. They built a fairly complex NLP processing system with a data exploration UI.

    I figured I had a group to help me figure out Python best approach in the space, but Python limitations just lead to stuff like needing a Kubernetes volume to host data.

    Conversely none of the data scientists we acquired were willing to code in anything but Python.

    I tried arguing in my company of the time there was a huge unsolved bit of market there (e.g. MLOP’s)

    Alas unless you can show profit on the first customer no business would invest. Which is why I am trying to start a business.



  • I don’t think its aged well.

    This game looked incredible for the time and introduced a rail gun sniper rifle you could one shot kill people with.

    This map let you camp out and be a sniper but it was possible to overwhelm the sniper so the game stayed fluid and teams had to support their sniper.

    Quake, HL2 Deathmatch, Counterstrike had similar weapons but quickly filled with people who could launch themselves 100ft in the air and headshot someone half a map away through a window which is why single shot weapons faded out of FPS games.

    If you try multiplayer on some of these games, the skill level of opponents is even higher, they know every trick and execute them flawlessly. This destroys the reason the map was so good.

    Playing the sniper in 4 player borderlands story is probably the closest you would get to the original experience.




  • Tesla actually market it as a positive.

    Car manufacturers have to setup different manufacturing lines to provide different feature levels. Tesla argue this makes them more expensive. Tesla cars have all features installed, just disabled and the optional extra packages are cheaper compared to their rivals as a result.

    To be honest there is a certain logic, if you’ve ever been in a Ford Focus LX (bottom range) its pretty clear they had to spend quite a bit of money on more basic systems. I honestly thought each LX was sold at a loss


  • I have a Mac book Pro for work.There is just a lot of random weirdness.

    There is no right click, your supposed to do a light two finger touch for right click.If you click too hard it opens the dictionary.

    If you plug in a mouse you can get right click, but it isn’t consistent in working.

    By default scroll is inverted (up is down, down is up), also windows can have scroll bars but they aren’t clickable, you have to do a scroll gesture.

    Almost every Left control + Button action is now Meta key + button. But not everything, its annoyingly inconsistent also new random shortcuts.

    For example lock screen isn’t Meta key + l like on Linux or Windows. Its Meta + Shift + Q, shut down is Meta +Left Control + Q.

    The keyboard doesn’t match the your countries layout, so keys move around and is missing traditional keys like print screen. To do that you press Meta + Shift + 4 to switch the mouse to a screen cut tool and select the area you cut.

    I could go on and on, none of it is obvious and I wouldn’t say any of it is an improvement at best its just different.


  • @ergoplato I didn’t suggest that.

    Personally I don’t think its ego. I think you have two issues.

    The first is people go through stages learning DevOps. Stage 1 has people deploy a CI because its cool, they build a few basic pipelines and then 90% of people get bored. The 2nd stage is people start extending those pipelines, it results in really complex pipelines requiring lots of unique changes based on the opinion of the writer. You move to the 3rd stage when your asked to recreate/extend for a new project and realise how specific your solutions are.

    Learning how to make minor tweaks and hook in a few key points to get what you want takes years. Without that most packagers will want to make big changes upstream which won’t go down well.

    The second issue, I have met quite a few developers who become highly stressed when the build system is doing something they haven’t needed to do or understand.

    A really simple example I have a Jenkins function which I tend to slip into release pipelines, it captures the release version and creates a version in Jira.

    I normally deploy it first as a test before a few other functions to automate various service management requirements.

    Its surprising how many devs will suddenly decide every problem (test failed, code failed review, sharepoint breaks, bad os update, etc…) is due to that function.

    For me this little function is a test, if the team don’t care I will work to integrate various bits. If they freak out, I’ll revert decide if it is worth walking them through the process or walk away.