I use WordPerfect 6.2 for DOS, not for any nostalgia or legacy reasons, just because it's a full-featured and highly configurable word processor that I can use in a terminal. I only use it for writing letters and so on, nothing too serious, but I prefer to stay in the terminal if I can.
It works beautifully under dosemu2, which has a terminal mode that can translate various VGA modes into S-Lang calls (S-Lang is like ncurses, so no X11 required). I find this technically impressive and makes a lot of old DOS software indistinguishable from native linux software; stdin/stdout, parameters, host filesystem access, etc all work transparently.
It can import TTF fonts and print to PostScript, which I just pipe into ps2pdf and then handle on the host.
I'm not aware of any other full-featured modern word processor that can run in an xterm. I know about wordgrinder but it's very very basic. You could use a text editor, but it's not ideal for layout because it doesn't understand things like proportional font geometries - you need that to know how lines/glyphs will fit on the physical page when it's printed. You could write it in some form of markup, html, TeX, markdown, whatever, but if I'm just trying to format a document I prefer a word processor.
(Note: dosemu2 doesn't require virtual 8086 mode, so it works fine on x86-64)
Brilliant - I've long observed that almost everything we are doing with our GHz computers, we were doing previously with old MHz systems. We have more pixels (but use them for the same result), more memory (but everything has grown to fill it), more bandwidth (but websites packed with junk bytes).
The only "new" thing enabled by our bigger systems is "big data", which is largely a process of finding patterns that will fool the user/purchaser/customer.
> I've long observed that almost everything we are doing with
> our GHz computers, we were doing previously with old MHz
> systems.
I do also, but then I do have to pinch myself and remember some of the cutting edge software that really does make use of the hardware - i.e. games, video editing, neural networks, etc. I know we could already do these things on lesser machines, but there is no doubt that the level of which they are currently done could not be replicated on a lesser machine. And also remember the efficiency of systems such as web servers, todays average usage would be a DoS attack of the past.
I do look in disappointment though at text editors, window managers, file viewers, etc - that despite having much more computing power, offer not many more features but still eat tonnes of resources.
I'm currently (slowly) working on a Window Manger for X11 which tries to bring back some of these ideas but for modern devices: https://github.com/danielbarry/oakwm/ It's built on top of wm2 (which itself has roots going back to Bell Labs plan9). A lot of the work is in ripping out the unnecessary features and making it touch friendly. The idea is to run it on the PineTab Linux tablet.
I'd like a browser that is just a nice browser. It should fit easily into a few 10s MB and start instantly. It wouldn't waste 100s of MB on implementing dangerous/pointless/wasteful javascript APIs, like battery state or physical screen APIs; it wouldn't support (literally) 4300 options; wouldn't save 200 MB of config data; and wouldn't load many MB of data just to start and display a blank page!
Dillo/NetSurf may be of interest to you. They fit the size requirement as well as the "lack of JS" requirement (i.e. no JS support at all). "appsites" will definitely not work in them, but the document-oriented web and things like HN are fine.
I've used Dillo, but unfortunately most web pages are seriously malformed. Would be nice if it was possible to get some better CSS support.
I would like to see better JS support, but the scope of JS is simply insane for such a small browser. It's unfortunate much of the web is completely unusable without running JS. Perhaps it's possible to first run the page through a larger browser engine and then send the processed content to the small browser (such as Dillo), that would massively widen the scope of what it could display.
For the record, chrome on android is ~60MB, less than half the size of desktop chrome (~150MB). Android chrome supports the same set of javascript APIs. Clearly that isn't the whole picture.
Actually, you can select which Chrome to use for the Webview in Android developer settings if you have Installed multiple versions of Chrome. Chrome stands on its own.
So Chrome for Android is 10MB larger than the entirety of Damn Small Linux, which includes a desktop environment with three web browsers (firefox, dillo and netrik).
While 60MB is still humongous, it does make sense that Google would at least care slightly on Android, as the majority of the Android market are low-to-mid-tier smartphones with reasonable performance instead of the "desktop-in-a-pocket" that are the current flagship phones.
But while they care slightly on Android, they don't care at all on desktops.
Depends largely on the state of code caches, microcode and bus structure. Loop unrolling can be the worst thing you can do, on some architectures.
See, the hardware folks have listened in on the compiler people and their problems. They've done things like identify loops and rewritten them in microcode for optimization. If a short loop can fit entirely within the CPU code buffer, speed goes way up. Unroll the loop and blow the CPU code buffer, defeat the optimization and lose all that.
what you want isn't possible with the web being like MTV and moving far beyond it's original intention. The best you can get is get one of the small browsers and live with broken pages. Some people do just that. I think you should be think in terms of relativity though. Think about how much a "10 MB" only browser took up in main memory back in the day.
> I think you should be think in terms of relativity though.
> Think about how much a "10 MB" only browser took up in main
> memory back in the day.
This is exactly the point, software has swelled to use the resources available to it, so with each new iteration your machine gets faster but what it runs gets slower. It doesn't feel like this is something we should settle for.
Well that and high quality video. One and only one tangible benefit of this new web is better quality video. Everything else sucks as much or worse than in the days of Altavista and punch the monkey ads.
But apart from the HD video, the search, the aqueducts, the roads, the education, the browsers that don't crash all the time, and the wine... what have the Romans ever done for us?
I disagree. Searching for keywords is practically impossible nowadays, with all smart AI-powered™ search engines that unhelpfully return results that are not relevant to your query. And SEO spam, which has killed the niche high signal-to-noise handcrafted websites for content aggregators.
Since 2-3 years, I saw a dramatic deterioration of search.
Whenever I am looking for anything that is not very mainstream, I get pages and pages of irrelevant results / SEO hacking sites
Remember "popup spam" and then how enough windows would open that none would respond?
I'm guessing anyone who didn't get browser crashes in the 90s was only browsing a couple of sites or something. Certainly anything before IE5.5 was crash city - in particular Netscape 2.0!
Chrome still crashes all the time for me - in fact, it seems to be crashing more than it used to. It's just that now the crash is isolated to a particular tab.
I've been a developer for a while, starting from QBasic, then Pascal, Delphi, C/C++, Flash, Qt, etc. happily switching from one technology to the next when it makes sense.
Today I use Electron and React Native, which needless to say are not very popular, but for sure I could never have developed the kind of cross-platform software I write today with the technologies I used many years ago. Partly for lack of skills, but also of time (cross-platform development was way more difficult), or simply because computers back them were not powerful enough.
I don't have any special nostalgia for old technologies, some stuff were good, some not so much. And as OP is showing you're still free to use old software if you don't like what's being done today.
MHz level stuff works for 'work' that normal people do - writing documents, emailing, etc. I dreamed for a while about creating a MHz-level processor from the ground up with an open hardware specification and creating a simple kernel + OS to run on top of it, with new secure protocols for networking (websites, not reinventing TCP/IP) and communications (email). Right now every single computer is backdoored by the governments of the most powerful countries: we have Intel MEs and AMD PSPs in the hardware, and the frameworks everything uses - Linux, OpenSSL, etc. - are almost all huge complex codebases that the NSA has tons of zero days for. If you could get open MHz-level hardware and get OpenBSD or something to run on it, and develop applications with a focus on secure code, it would be pretty great.
But ultimately this would all fail. Aside from the network effects - nobody's going to use it - the fast processors and tons of memory are necessary. We just don't think about it unless we're running out of it.
- high resolution photo and video - photo editing programs like GIMP and Darktable can spend some time processing photos; these days even with complex effects you usually never experience lag more than a few hundred ms on many megapixel photos, because our hardware is fast. Same for video, the memory and storage space and bandwith is a hard requirement, and going back to 360p is not really acceptable.
- high resolution monitors - no point in having great looking 4k 10 bit color video without a 4k monitor, and now you're stuck having to push 20gbps through your displayport cable on a 500 MHz processor. And text also looks much better at 4k. Also you could say this is a bit unnecessary but compositing window managers are pretty great and I would say a core requirement of modern GUIs, and having eye candy like wobbly windows, window shadows (those are actually very helpful, try turning them off), etc. is expensive. My Dell XPS laptop from 2017 couldn't handle wobbly windows without visible stutter at 4k60; my desktop with a $400 GPU from 2019 can keep up at 4k@144hz.
- new technologies - VR/AR, fast voice recognition and neural networks - this is all cutting edge stuff but the use cases are obvious and they have started to be applied more commonly. Also the failures like eye tracking, Kinect - they may have failed commercially but they were good ideas and a valid use of fast computers. Also online meetings with many participants, each with their own video streams that require decompression.
- obviously, video games - not that you can't have fun games with shitty graphics, but good graphics are nice.
I thought of a few more but can't remember them at the moment. Oh, also, Bluetooth audio - 192kbps audio, but you need to decode it since it takes compression to get it to that level, and then you have the additional overhead of sending it over a digital protocol instead of just having your ADC do the work. You would need expansion cards or a dedicated audio core in your CPU to accomplish this if your computer wasn't fast.
And of course, science needs big data and big hardware to process all of that data.
If you're only working with text then slow computers are fine, but as soon as media gets involved they are not practical.
Don't get me wrong - I like hi-res images, and bluetooth (when it works), and I'm still amazed that my phone has more grunt that a Cray.
But the stuff we do day-to-day (excepting graphics) is appallingly inefficient. And most of the look-at-this stuff (speech recognition, and you mention VR and eye-tracking) is not actually used for anything day-to-day; we don't routinely talk to our computers ("Hello Computer" https://youtu.be/v9kTVZiJ3Uc?t=10 ).
That's perhaps because doing so in an office or cafe environment isn't great.
Kids are using text-to-speech more while doing their homework. Word has been beefing up its transcription feature for a while now[1].
I got to see this happen during lockdown, where teachers I know recommended transcription for younger kids who don't have touch typing skills and were still expected to turn in work on a computer. Talking to your computer is much more natural while doing your homework in your room.
When I was a young person with executive function issues (late 90s/early 2000s) voice to text was a lifesaver for my ability to write coherently. I found it much more natural (even then) to "write" by speaking my mind to a computer than to struggle to focus while typing slowly. It wasn't until I was older and using IM heavily as a social outlet (which was much more in the "bursty" configuration that works well for my brain) that my typing ability caught up with my pace of dictation.
Saddest part of speech recognition is with all the gigahertz and storage we have, we send it out to the internet to be processed by something else then get the result.
> Right now every single computer is backdoored by the governments of the most powerful countries: we have Intel MEs and AMD PSPs in the hardware, and the frameworks everything uses - Linux, OpenSSL, etc. - are almost all huge complex codebases that the NSA has tons of zero days for.
You've giving the gubment too much credit. It's easy to throw up your hands and say "aww geez I guess everything is backdoored anyway, why bother?" -- this is exactly what they want. The truth is a lot more complex, and as a lot of leaks have revealed, their capabilities are far from the supernatural omnipotence you seem to be implying.
Not perceived: actual, effective latency is higher. Today's hardware is capable of better latencies, but most software guys don't care about latency all the way from firmware to OS to apps to websites. It does matter to some degree in games and vehicles but it's hard to find an industry that values latency outside of toys or industrial applications. Practitioners say this is because "99% of users don't care" but that's only because 99% of users have either (1) never experienced fast latencies or (2) have forgotten them because latencies have gotten gradually slower over the last three decades.
All large systems of systems have "duct tape", I'd go as far as to say it's an emergent property of systems of systems.
It comes about because unless you design every sub-component system in lockstep with every other (which is impossible, the romans didn't lay out londons for cars, the victorians didn't put in sewers with respect to where we'd want to run fiber) you end up with an impedance mismatch at the boundaries.
It's why back-hoe operators from a gas network rip up fiber depressingly often, why (in the UK) the UK transport system has choke points between road, rail and air and on and on.
Modern computers aren't a unified system, they are lots of seperate systems that talk to each other and frankly having some (minimal understanding) of what has to happen for Gnome to appear on my screen when I press my power button I'm amazed that it ever works never mind mostly without fuss.
Exactly. Even if you do care, you can only exert so much influence on the platform you're developing for, unless you're working for a company like Apple or Sony. Even then, there are many, many layers to the latency problem.
Our systems are doing so much now, and all that "extra" lets us do numerous things at once without blinking an eye and it allows new comers to step up to a computer and almost immediately start using it after some exploring. Not clicking something and waiting 20-30 seconds to see what the program is or a blinking green cursor wondering what to do next.
> Sure you might be able to do some of those things at a very low resolution on a Mhz system, but not in any meaningful way.
> You must be really young.
Maybe they are, maybe they aren't, but that's an unfounded assumption. There are plenty of middle-aged people who haven't got a solid idea of the limits of pre-Ghz-scale hardware for use cases that weren't all that mainstream back then, and I'm sure there will also be young people who've tinkered with some old Windows 98 box and could give us a pretty solid opinion.
At any rate, I've used a succession of Mhz-scale systems (starting with a 386 that struggled valiantly to run a copy of Windows 3.1 that someone had installed on it) and I'm definitely not very young.
> >* Graphics design
> We did
With rough and simple tools, limited to low-res designs, very limited brushes, super crude brush simulations (if any). Procreate on the iPad blows all of it out of the water effortlessly with its worst brushes using a finger for a stylus, and that realtime brush simulation has been taxing for earlier iPads, that have always been orders of magnitude above Mhz systems. I used Photoshop CS and CS2 a lot back in the day; they were resource hogs and still very crude compared to current entry-level apps. We've really gained a huge lot in that department.
> >* Video editing
> Ditto
At what, 640x480, 15fps? I remember having annoying render times for things that wouldn't hardly even count as filters nowadays at such resolutions, maybe 800x600, but I'm sure that ball-shaped logitech webcam could do no more than that. Snapchat replaces whole faces in realtime with eerie accuracy on what must be much higher res.
Color grading Full HD, 30fps? As if. I guess you couldn't even play that without stutters unless with dedicated silicon to begin with.
> >* Animation
> Ditto
Low-res, crude, and anything halfway detailed and interactive: Low-fps. Manipulating anatomy models like those at https://www.zygotebody.com/ in realtime? I doubt my Core 2 Duo laptop would have been up to that. I had similar software for Windows 98, and it was absolutely primitive and still taxing the CPU and GPU.
> >* 3D CAD
> Older than you think
With less precision, much, much lower model complexity, and much cruder tooling. A large number of components, complex shapes, joints, ... that's going to hit hard limits very quickly. I'm hitting hard limits with that sort of thing nowadays, but I'm hitting them somewhat later than even a few years ago.
> >* 3D Rendering
> 3D rendering was born here, and we were playing 3D games perfectly in 1996-2002
I don't play a lot, but between Gothic 1 and Witcher 3, graphics have improved by an incredible amount, it's day and night, and I can't even go to full details in Witcher due to my aging GPU. Technically, those systems could do it, sure – but only with super short visibility, very crude models and extremely limited shaders and animations, crude collision detection, ... Gothic required at least a 700mhz Pentium 3, so it's pretty representative I think. Of course, those limitations work better for some games than for others, but they still were brutal limitations.
> >* Multitasking
> Linux, KDE3.
Also quite limited, though. Just what fits on a single Mhz-scale core. On my Windows 98 PC, I had to close CPU-intensive applications all the time because things would start to stutter; I believe Windows 95 would sometimes just bluescreen under such conditions. Things got a lot better on Windows 2000, but I think that may have been on my 1Ghz Athlon already. Those early Linux desktops were pretty unstable with lots of multitasking as well, I faintly remember lots of freezing. Things did slow down perceptibly at any rate when doing multiple resource-intensive things. Technically possible, sure – but it absolutely helps to have lots of fast cores.
Of course, a lot of the legwork to make those use cases perform is nowadays done by GPUs or huge amounts of RAM, and profit lots from multiple cores, but I'd say a Mhz-scale system should have a period video card, too, and MB-scale RAM, and be single-core, otherwise it's kinda pointless. And under those conditions, all of above things were technically possible, but really severely limited – still are, in some cases (CAD...), but wayyy less than back then.
Does that qualify as "not in any meaningful way"? I guess it depends. It was meaningful for me back then, and it made possible things that hadn't been possible before, and of course we were always content with what we had, not like thouse young'uns nowadays, and walked uphill to school, in the snow, both ways, every day – but looking back, the capabilities of my Mhz computers feel incredibly crude and primitive by today's standards, and even a lower-end gaming PC has little problem running Autodesk Fusion 360 (for which there's even a free hobbyist license) with models of surprising complexity, and I'm sure that enables many many more things that wouldn't have been feasible on Mhz hardware.
TBH w98 was a joke on multitasking. You can't even compare NT4 or W2000 (not mention to Linux) with Windows 98, where it struggled even under a Pentium 4.
On freezing, Mandrake was a joke, but Slackware and Debian were rock solid.
My parents got their first computer in the 90s so my mom could do freelance work editing medical texts on nights and weekends after her day job. She used Word Perfect and she taught me to type with it, I still remember the exact shade of blue and the white text. She’d use Reveal Codes and I thought all the special characters were so cool. I can still picture her sitting at the desk working on it while I’d play video games. I was maybe 7 or 8 years old. They’re vivid memories that are kind of meaningful that I haven’t thought about in decades, tied directly to that application. Weird how that works.
Somehow WordPerfect/MultiMate with reveal codes is still a more intuitive text formatting tool than anything I've used since. Or maybe it's just nostalgia. There's nothing I hate more these days than trying to format text in WYSIWYG editors, in-browser confluence and jira, notepads, markdown, etc.
I wish there was a mandate that companies had to register sourcecode, e.g. with the Library of Congress, so that it could be released as Public Domain after a couple of decades. We are losing quite a lot of our software heritage when companies bury projects like WordPerfect or milk older games for a few more bucks. We could have a thriving, open-source ecosystem around the sometimes excellent software that is currently locked away.
We've decided 20 years is long enough for drug patents, and in my opinion, it should be good enough for almost all intellectual property, really.
Create something new? Great! You're officially granted a monopoly on that thing for 2 decades - plenty of time to monetize it, and use the money to create some other new things. If you can't figure out a way to monetize it in that timeframe, then we'll allow others to compete with you and take a crack at it.
Given that Disney made its fortune by adapting public domain stories and then abusing the copyright protection of their adaptation of these stories, I think there are more loopholes to patch than just the expiration date.
Very cool! I couldn't find a trove where I could browse by category or other metadata, although I did find the search interface. Is there a categorised trove?
I had no idea dosemu2 could translate VGA to S-Lang. That's very nice, and I will look into it.
WordPerfect was a great word processor, I remember using version 5.1 for years because it was wicked fast, rock stable, and most importantly, it was predictable. By contrast, MS Word -- even today -- seems to have a mind of its own. You move an image, and suddenly all your numbered bullet points appear in a different font except for the last one, and nothing short of retyping the whole thing in another document seems to fix it.
That’s a problem that no WYSIWYG has been able to solve yet. It’s probably an unavoidable abstraction leakage.
Hence the unbelievable popularity of markdown.
Imagine telling your 90s self your current machine configuration and how awesome it is. Then explain that we, with all this tech, voluntarily chose to type in plaintext format, with notations that kind of work as a more readable form of markup language, because we got fed up trying to make WYSIWYG work.
Oh yes, and GUIs are for noobs, pros use terminal emulators, but that’s another topic :)
> That’s a problem that no WYSIWYG has been able to solve yet. It’s probably an unavoidable abstraction leakage.
I think it comes from conflating layout with writing. In QuarkXPress or InDesign it really isn't an issue. Those tools are quite uncomfortable for sitting to write in, though.
LyX does a decent job of this, I think, though it's not exactly WYSIWYG.
Have you actually used the GUI version of WordPerfect?
I was forced to use it heavily for a year of pure writing for a previous job. It's incredibly unpredictable. It's like whack-a-mole. You make text in one place bold, and all the sudden some of your footnote text on a different page becomes bold.
The people at that job who were really good knew all of the tricks, digging into the codes that WordPerfect inserts to address various issues. But even then, it was an extra step, and I never became as productive in WordPerfect as I had been in Word.
Plus WordPerfect has been on maintenance mode with Corel for decades at this point, catching lots of bugs and half-implemented features.
Replying to say you are correct and GUI/Windows WordPerfect was completely 100% broken when dealing with 'closing tags'. Some people knew & liked to use the codes, but the basic GUI editing was just defective and caused your formatting to spill out randomly. (at least in vs 5 & 6)
Obviously I get this with HTML, but the Mac & MS Word approach of 'object oriented formatting' was just a much better execution for mouse operation.
The only version of WordPerfect I ever used was 5.1 for MS-DOS, and my comments apply to that one only. I heard the GUI version left a lot to be desired, so I never upgraded.
> You make text in one place bold, and all the sudden some of your footnote text on a different page becomes bold.
Tbqh, this is my experience in just about every rich text editor. Well, maybe not quite that bad, but I’ll always bold a word, and then while editing I have to go back and change the word after it, and it will suddenly be bold.
I wish bold/italics/etc acted like caps lock. It’s either on or it’s off, and the computer doesn’t try to guess for me.
Outlook has a setting to automatically capitalize the beginnings of sentences. I turned that off. But it's been long ago, I forgot where that setting is. But I recall it was near the setting that enables smart quotes.
They even do it on the web version of Outlook. The setting to change it is 4 levels deep and two of the options are at the bottom of lists and labelled "more options".
On display? I eventually had to go down to the cellar to find them... It was on display in the bottom of a locked filing cabinet stuck in a disused lavatory with a sign on the door saying "Beware of the Leopard."
I haven't seen it with Word, which I use quite a lot, but there is one similar and extremely long-standing bug. When you use a keyboard shortcut to turn bold/italics on, then type a word immediately before a non-whitespace character, then use the shortcut to turn it back off, it will un-bold the thing you just bolded. So annoying.
I am having flashbacks to late 90s Linux. One of the distros back then was Caldera OpenLinux. It had a Linux port of WordPerfect. It got a lot of criticism at the time for being nonfree.
It requires libc5 - the c library that was common before glibc became common at the end of the 90s. And then needs X libraries that are compiled against libc5.
Probably you could find an old Red Hat 5 or something and pull out those libraries to run on a recent kernel.
I think WordPerfect 8 was where Corel rewrote it. It didn't work nearly as well. I vaguely recall they switched from native to some kind of compatibility layer (maybe Wine?) which just wasn't very good.
I wonder if libc5 is so old that it has problems with docker. But I remember loki games being statically linked to avoid problems with dependencies. You only needed like linux syscalls and X11 for them to work.
docker has no concept of the binaries that actually run, it just sets up namespaces, cgroups and a userspace. worse case scenario, you have to run the docker container as privileged, still no difference from running on host (at least with a properly constructed image, where the binaries wouldn't run as root)
Nota Bene is another text-based word processor from that era. It defaults to showing embedded formatting commands. The primary means of operation is via an integrated DOS-like command prompt.
Nota Bene although no longer text based still exists and is actively developed. Nota Bene used the XyWrite word processor engine. XyWrite was very popular in news rooms back in the day.
That's a description that also matches Arnor's Protext. The command window was the bottom half of the screen, and one embedded rulers and suchlike into the document text with ">" lines.
This is testing my memory but I think WordStar from the early to mid 80s had something similar. Codes all started with a ^ (again from memory) and their visibility could be toggled on and off.
This week I built a home-made computer using an ESP32 instead of a Z80 and hodge-podge of 74xx logic chips to run CP/M and started playing with Wordstar 3 and 4. I used it many decades ago briefly on the Amiga and it was more graphical but I'm only just getting to grips with it now.
I have known lawyers and legal secretaries who just would not give up on WordPerfect, I think because of the footnotes. The two that I most recently had dealings with have since retired, but I'd be anything but surprised to learn that their are recent .wpd (is it?) files on the General Counsel's Office's shared drive.
I liked WordPerfect well enough from 4.2 through 5.1. Version 6 on Windows was not good. Even a slow typist (me) could get ahead of the cursor. But at some point my employer switched to Word and I haven't been back.
During the 4.X and 5.X era at WP, the company paid very close attention to the needs of the legal market and WP could do things with a couple of keystrokes that would take 10 minutes with MS Word.
WP also had a potent macro capability and a good macro editor as well. I was practicing law at this time and had created 80-99 WP macros that made legal docs appear like magic.
There was no e-filing and court clerks and judges could be more than a little picky about the formatting of the paper documents they received. If a doc wasn’t formatted in the proper manner, some clerks would refuse to file it and return the doc to the lawyer’s office for a re-do.
The acquisition of WP by Novell was a disaster. Way different corporate cultures and leadership styles.
Ashton and Bastian had been generous with stock grants to key people, particularly on the technical side. Most of the key people below the two founders had enough money that they didn’t have to put up with the sharp elbows and confrontational styles that predominated at Novell.
Combine a brain drain with a Microsoft push for Windows and built-in integration with the first passable version of Word and you had a recipe for disaster. Some of the WP people who remained claimed that MS had mislead them about Windows vs DOS roadmap and been slow in providing Windows info necessary for WP engineers to build a decent first release of WP for the new release of Windows, but I could never determine if this was what had really happened or just sour grapes because WP stumbled badly after the ownership change.
It was real. They built their own language and compiler to compile to undocumented windows API calls because MS wouldn't share in the early days while they established a WYSIWYG foothold.
In the corel days they finally decided to port to actual win32 code and put aside the custom transpilation
According to this history written by one of the founders they were given a pre-release copy of Windows to get an early heads up on porting, but the engineers generally preferred to work with DOS:
> I'm not aware of any other full-featured modern word processor that can run in an xterm
Word Perfect was available for UNIX as well. The SCO binaries might run with iBCS. I don't know if that would provide a significantly different experience than running the DOS version in dosemu though.
I tried running WP for SCO under iBCS 10-15 years ago. I couldn't get it to work. Don't remember all the details but it seemed they cheated with some of the video routines and I mostly got a corrupted screen. I didn't spend a lot of time on it, though, and I very well might have just not hit on the right incantations.
Doesn't RR Martin use the original Wordstar? I remember reading about him using an old school physical word processor and USB-floppy drive when he needs to backup/transfer his work.
Version 4 got me through the first couple years of college. My freshman year I actually had a steady stream of people coming into my dorm room to use it on my PS/2. Eventually I switched from English to Math and moved on to vi and the various roff markup tools and then LaTeX, but I think what ultimately killed the magic of that generation of WordPerfect was the change in keyboard layouts that moved the function keys from the left side to the top. Before that when combined with Ctrl, Alt, and Shift there were 40 commands available just a pinky reach away, and it came with a cheat sheet card that fit neatly around them.
Versions 4.1 and 4.2 got me through the first couple of years of college also. I briefly updated to 5.0 but quickly reverted to version 4.2. I have never been as productive in a word processor as I was using WordPerfect and a Model F XT keyboard.
IBM spun out their keyboard division decades back when they they spun out printers (into Lexmark). The current inheritor of the keyboards is Unicomp (having bought it from Lexmark/IBM): https://www.pckeyboard.com/
It's a bit cheaper to buy from them than to pay for an antique on eBay.
Also, there's a wave of mechanical options from more recent companies like Das Keyboard or the various DIY kits with choose your Cherry switch adventures.
Reveal codes should be part of any rich text editor. Ever sat wondering why pressing return added a new bullet to a list rather than a gap to the text preceding it? I see it all the time while watching engineers and manager using tools like Jira, Confluence etc. Reveal codes would make it plainly obvious - the cursor was after the list start but before first bullet.
Nothing more delightful to work with than a requirements management tool whose wysiwyg editor changed the text of the bullet points when the indentation level changed. Never figured out what was going on, but there was somehow two different texts occupying the same line, and if you pressed tab to make the line deeper in the hierarchy it would display the other text. You start wondering how bad DOORS really is if this is how the "modern" web app replacement works.
I really don't want a word processor, per se, so much as I want really robust (realtime) text wrapping/unwrapping as I edit paragraphs. I've tried to get Vim and Emacs to do reflows but both were very flakey compared to something like WordPerfect for DOS.
I am finding that semantic HTML and CSS Grid where you style the elements works nicely for documents. This goes against what Tim Berners Lee imagined in that I am typing in the HTML tags and happy to navigate them in text.
I prefer the tags as they describe my work. Visually there is no evidence that something is a 'section' or an 'aside' but I make sure my words do have document structure, so a 'section' will probably have a 'h3' at the start of it.
I wish there was a 'what you see is what you mean' editor that was visual but only allowed you to add HTML semantic elements. WYSIWYG editors tend to just add local styles and 'br' elements, resulting in HTML that is useless for my style of HTML.
I can do diagrams in various ways including SVG that I just write out manually. Or I can use the wonders of CSS Grid and a 'dl' list with first of type/last of type CSS selectors to style up a simple linear flow diagram.
I really wish there was a super neat word processor or HTML editor that only wrote lean HTML with semantic markup for document structure. But here I am in 2020 using my IDE or Vim to write documents and even craft diagrams.
I am impressed that I am not alone in spurning drag and drop wonder gadgets. As much as I would like to give Wordprerfect a go, I feel the world has moved on from documents printed out as memos and put in pigeon holes. Stuff has to look good in a browser and with stripped down HTML that uses the elements and a simple stylesheet that loads a font and puts in some CSS Grid layout I feel I have the tools I need.
Word perfect 5 and 6 in many respects are still vastly superior to any version of word, especially when it comes to aligning text left and right, which are properly treated as line flow instead of a properties of a tab stop.
Without leaving the keyboard I could easily align text on the same line to left aligned, centered and right aligned.
For typing / journaling the non-wysiwyg was a feature. I honestly think wordperfect 5.1 is a big reason why sublime is so popular now.
Also I'd use WP 6.1 for DOS in Spanish if it wasn't a "little" inconvenience: we don't use Pesetas as a currency anymore, but I think there's an Euro patch somewere.
EDIT:
On graphics, that's a solved problem. Everyone uses framebuffer today, even on obsolete chipsets it's doable. Just map VGA calls to FB and everything would be fast enough, I think.
I used WordPerfect in my first job as a tech writer. It was beneficial to have a separation between writing and the layout done by desktop publishing systems.
As a programmer building websites, I do the same thing, writing in vim. Learning how to make your editor work for you is a great investment.
What macros do you use? I have some simple programs that generate documents to print out, and I use LaTeX and a bunch of packages. It takes hundreds of megabytes of software with all the packages. I've long toyed with using the MOM macros but can't get motivated to rewrite my programs and learn troff.
It used to bother me that groff doesn't support Unicode (as far as I can tell) but then realized that all I write is English so why am I fetishizing Unicode? groff will get me any accented character that I would realistically need.
You can have UTF-8 with groff using the preconv [1] tool, just pass the "-k" or "-Kutf8" paramater. It will preprocess your source and replace Unicode symbols with its groff special character form. Of course, the built-in standard fonts do not support many glyphs but it is no big thing to use a more Unicode friendly font like DejaVu Sans or Noto, check [2].
So this is my setup using MOM macros for nice typesetting.
But I fully agree on your last point, in the end I seldom need the Unicode universe for most of my documents.
Little-known fact: groff (more particularly grotty) does colour, albeit that people have secretly turned this off for you. grotty's own man page is in fact in colour.
I have very fond memories of WP5.1.... I still have some Model-M keyboards, so I could really recreate the "glory days" if I wanted. I could print out the little F-key cheat sheet for the keyboard as well.
But...
What about Unicode? Lack of Unicode seems like a dealbreaker for a lot of use cases. Even if you're not using Unicode, seems like you'd eventually want to work with Unicode text.
As an engineer, a lot of the "word processing" I do involves cutting and pasting text from other sources (citations, code samples, names, etc) that tend to be chock full of non-ASCII characters.
Or has this been solved somehow? I see that there are various hacks involved in getting support for the Euro symbol. What about a general purposes solution? I know that WP5.1 had support for international character sets. Perhaps somebody's cooked up an emulation layer that translates Unicode into whatever it is that WP needs.
It's also an odd little page. It includes directions to buy it, if you already have a copy. If you don't have a copy, it includes some details on how to use eBay.
Former employee had a Inventory system based on dos6 but the hardware (the pc and the barcode scanner) was getting really old. So i buy a new PC, installed FreeDOS and a new barcode scanner with Dos compatibility and it works perfect. DOS especially FreeDOS is just great.
Wow. I think Wordperfect may be the last time I enjoyed using a word processor. It was also maybe the first time I realized how important good design is. I couldn't you tell a specific reason it was better than Word, just that it always did what I expected.
Like your style on this. I'm contemplating similar setup but going bit more oldschool with Word 6.0 Win 3.1 edition, gotta work out what the smoothest way of integration with my Linux (Mint) desktop though. Integration is key for me and have to say my prior exp with Wine have been less than comforting (working but just so, not long term functional feel).
Wine doesn,t work great with Microsoft software, because they would take advantage of undocumented APIs, but W6W31 may be old and basic enough to work...
Nice that you are using WP on dosemu2 on Ubuntu on Windows. I once spoke to someone active in research into keeping data (formats) and programs available for very long time (like centuries) and I always thought it was a no-brainer because one could just run X on Y on Z on A on B on C to get ancient software D working.
Allthough I also have to admit that this approach didn't work for my favorite game [0] but now that I'm googling, I does seem to work! So someone got DirectX 9 or something working?
So like SLRN and Jed? Amazing. I used DOSEmu back in the day for Redneck Rampage under an AMD Athlon (Dear DOSBox users, DOSEmu was zillions times faster than DOSBox, I ran it at native speeds). This is impressive.
While I support your travels with Wordperfect (I wrote many term papers on that puppy), I can support the quote by the person you responded on twitter. He just showed arrogance and silliness when he said
"All that was ever really needed by an OS or Office apps was already there 15 years ago. Everything "invented" beyond that was just vanity changes with negligible long term added value, and constant moving around of UI to appear new and better. Change my mind."
I was a huge fan of Wordstar 6! Wrote so much with that thing ... way better than Word, Pages etc. even still. It's no wonder I prefer Markdown for writing / documents etc.
I should try to see if I can get Wordstar running under dosemu2 (I probably have a 3.5" floppy somewhere with it on, but I don't think I have the disk drive to read it).
Amazing. I used WP6.2 to write papers in high school and remember not only the tables and line drawings but you could also place images and annotate them. It's hard to remember how that worked but in those days it seemed those things were only possible on Macs. As I recall there was a big reference binder that documented the features.
Most of what I do in a word processor today, I could have done in Wordstar 3 under CP/M. Perhaps only the printing (no postscript) would have been a problem. Certainly most of what I do in Excel could be done in SuperCalc.
I remember using WordPerfect on VAX/VMS (on a decterm, no less), and I recall them having some flavor of Unix based binary, but, I can't find references to either out there anymore.
There is Wordgrinder[0] and can export to LaTeX for printing.
In classic Unix tradition, it looks way more plain than the screenshots of WordPerfect you posted :-P.
I had that version of WordPerfect on my 286! I remember how you could install the Trident video drivers and get some pretty high console resolutions. F11/Reveal Codes was super helpful too. I wish modern word processors had something similar. It basically shows you the equivalent of formatting tags.
I have the Win3.1 version of WP from that era in a DosBox. I've been meaning to compare it's grammar check to modern word processors. It was way ahead of its time.
I know Corel bough it, but it looks like it spun off again. I wonder if the current release still has any of the old source code, or if it was just rewritten and has the name for the brand: https://www.wordperfect.com/en/
Still owned by Corel (they have a weird "Corel Advantage" logo at the bottom of that marketing site and the Contact Us link goes to a bunch of Corel info).
The Corel of 2020 seems a fascinating Enterprise Licensing zombie. They appear to make all their revenue from (very) legacy Enterprise Licenses and government contracts and all of the software seems stuck in a time warp stasis with only the bare minimum of maintenance (presumably just enough to keep the Enterprise Licenses alive and well). It's almost sad to see how many classic computing brands they own and let languish in this state.
[1] https://www.bbc.com/news/technology-27407502