I love that they specify that they’re not accepting pull requests.
The MS-DOS v1.25 and v2.0 files were originally shared at the Computer History Museum on March 25th, 2014 and are being (re)published in this repo to make them easier to find[.]
In 2014, MS-DOS 1.25 and 2.0 were released under a Microsoft shared-source license (Microsoft Research License) which forbids redistribution
In 2018, both versions were published to GitHub and relicensed as MIT, making them properly open-source
Today, MS-DOS 4.00 was added to that repo, also under MIT.
Ignore them. Send a pull request with the full source of Arch Linux.
Nah, just a giant compiled binary blob. That’s what all the cool hackers do these days.
dumb question maybe, but where is the full source of arch Linux? My understanding is that its just vanilla Linux that uses the pacman package manager.
Am I wrong in saying the pacman is the Arch source? Or is there more going on in the tar ball?
Cheers! It looks like this is then the PKGBUILD
https://gitlab.archlinux.org/archlinux/packaging/packages/base/-/blob/main/PKGBUILD?ref_type=heads
In which case, there are no packages defined there which are Arch specific except pacman. So… pacman is the Arch source, right?
There is even a sentence in README.md
that makes it explicit:
The source files in this repo are for historical reference and will be kept static, so please don’t send Pull Requests suggesting any modifications to the source files […]
LOL, some of the comments in the source are gold.
https://github.com/microsoft/MS-DOS/blob/main/v4.0/src/DOS/ABORT.ASM
Note: We do need to explicitly close FCBs. Reasons are as follows: If we
; are running in the no-sharing no-network environment, we are simulating the
; 2.0 world and thus if the user doesn't close the file, that is his problem
; BUT... the cache remains in a state with garbage that may be reused by the
; next process. We scan the set and blast the ref counts of the FCBs we own.
;
; If sharing is loaded, then the following call to close process will
; correctly close all FCBs. We will then need to walk the list AFTER here.
;
; Finally, the following call to NET_Abort will cause an EOP to be sent to all
; known network resources. These resources are then responsible for cleaning
; up after this process.
;
; Sleazy, eh?~
i remember writing .bat files and pretending they were really fancy update scripts when i was like ten they did nothing but it was still fun :)
Like half of my job is writing .bat files to automate stuff locally and not tell my boss that all I do anymore is double click the right things in the right order…
You can put in a timeout command at the end, and then call the next .bat file.
For example “TIMEOUT /T 60” waits for 60 seconds before resuming, or you can override it by pressing any key.
So if you know how long the wait time between scripts is, just write a master.bat and call them in order, with adequate waiting time in between.
Okay so the dumb part is a lot of this is me abstracting away our complex build system. I’ve basically bubble-gummed a dedicated build system in top of it for only the tasks I do. At a certain point if I start adding configurations or timing I might as well just wrap it in gradle or something. But the system that I’m calling is already their attempt at simplifying another build system that’s underneath it that was written by the old guard using arcane sorcery. The whole thing is a mess
Next step, bind them to unused keys on your keyboard and press them in the right order
i propose to create a menu in which you can define what batchfiles to run in what order. its been a while since i worked with batch files, but if memory serves right, that should be doable, no?
only if they keep the it room as dark as possible and whenever someone walks in for help they jump up onto the desk with a flashlight and yell "I’M BATMAN’ while wearing a cape and underpants.
I had a job like that and powershell was a godsend. I let it slip when I accidentally set the multiplier for the delay randomiser too low and it did a months work in a morning. I ended up writing a guide for the others there when I left but sadly everyone but me had computers that supported newer versions of Windows where the scripts ended up broken. They asked me to come back and update it the Monday after I left. I asked if they would pay me to do it. They said no. Then I said no.
This is part of why I still have manual kickoffs for mine. Never need to worry about work getting done while I’m away or getting done suspiciously fast. Also they should have paid you lol, the dingdongs. Would cost a lot more just in work lost having someone else spend time deciphering and fixing it. They could always get someone else up to speed with the system after it is fixed by you so there’s little or no down time
They were important to boot games that needed most of your limited memory.