Steve Jobs, NeXTSTEP, and early object-oriented programming (2016)

1 year ago (computerhistory.org)

I am not an Apple user but I do enjoy the history of computers -- which (obviously) the timeline will reach the likes of Steve Jobs and Apple. As much as people give high praise to Steve and the rise of Apple... I personally value Steve more regarding his work on NeXT more.

To be honest, as someone who was into SGI computers as a kid, accepting the reality I would NEVER be able to own one, I hoped one day I would get to use or try one in the workplace. Of course, by then, technology had moved and these expensive machines are now affordable, simple a Windows PC with a good graphics card!

Getting back to NeXT, I heard "Next" thrown about when reading up on ID Software (Doom, Quake, etc) but never really appreciated what this was until later on. When I finally spent the time to understand that this "Next" was a computer, I started to appreciate how cool it was especially back in 1990! It made me appreciate how far ahead they were with programming/development tools. I mean.. Visual Studio finally caught up with productivity.. like.. 10 years later!

It is great to see the history of NeXT continue -- not just being "merged" into Apple but with OpenSTEP and others.

It also makes me appreciate less known programming languages that, one day, could become "mainstream" at any point. Hardly anyone knew much about Objective-C but became a popular tool for Apple developers with OSX. Thats why I always have a grin when people mock a less known language because "hardly anyone uses it" or "hardly anyone knows about it" -- that could change, you know.

  • But the RAM requirements were crazy, 8MB RAM in the late 80s which is understandable for what the machine the was offering.

    Interface Builder is nice, I remember trying to develop GUIs on Amiga with intuition.library and wished that there was a drag and drop program to build the source code. Running NeXTSTEP 3.3 on Virtual Box and trying to get it running on my early 2000s laptop :)

    • Now imagine being at the university doing the graduation project porting a visualisation framework from NeXTSTEP into Windows (Objective-C => C++), because it was clear NeXT wasn't going to survive, and some stations were already gathering dust, after being replaced with PCs running Red-Hat Linux.

      If folks had guess what turnaroud NeXT would go through, most likely my thesis would have been something else.

Related:

- Oral History of Blaine Garst [PDF]: https://archive.computerhistory.org/resources/access/text/20...

- Oral History of Blaine Garst [Video]: https://www.youtube.com/watch?v=qtEIq7fe_KQ

- Oral History of Steve Naroff [PDF]: https://archive.computerhistory.org/resources/access/text/20...

- Oral History of Steve Naroff [Video]: https://www.youtube.com/watch?v=ljx0Zh7eidE

  • I highly recommend the Steve Naroff interview. It's 6 hours in two parts, but it's worth the investment.

    Not listed here, but also available, is the oral history of Brad Cox and the later Objective-C paper authored by Naroff, Cox, Hsu (the author of the linked blog post) for HOPL IV. Someone else will have to dig up the links.

    You can also find interviews by other NeXTSteppers on the CHM site like Avi Tevanian and others. I think the Naroff one is best.

    • I always found Objective-C really elegant, due to its Smalltalk influence. I once read about Brad Cox explaining that the initial prototype was just one or two pages of C code to add objects and messaging. It was interesting to discover that Erlang had a similar origin story, with a thin layer built on top of Prolog [1]. Does a similar description of early Objective-C exist? I couldn't find a pointer in the videos.

      [1] https://www.labouseur.com/courses/erlang/history-of-erlang-a....

      1 reply →

    • Thanks for mentioning this! I watched the first part and just started on part 2. Very interesting!

> By the 1980s, the problems of cost and complexity in software remained unsolved.

We’re well into the 2020s and I’d say those problems have pretty much remained unsolved. ;)

Ah, Objective C. The language that combines the famous memory safety of C with the blazing-fast performance of Smalltalk.

Why do the words "Steve Jobs" and "programming" in the same sentence very slightly rub me the wrong way. Probably because one thing he wasn't, amongst all the superhuman things attributed to him by Apple zealots, is a programmer.

  • Steve Jobs was first a computer hobbyist, then later he worked at Atari before he co-founded Apple in 1976. He wasn’t a non-technical founder, although he was certainly less technical than his co-founder. He could program, he could do digital design, he knew his way around a bread board and an oscilloscope.

    He isn’t famous for being a programmer, but he could do it, and he certainly understood it.

  • He actually did know how to code. He and Woz created the classic video game Breakout while working at Atari. (I have no doubt Woz did most of the heavy lifting.) It's not that he didn't know how to program, it's just that coding was not his strong suit.

    • > He actually did know how to code. He and Woz created the classic video game Breakout while working at Atari. (I have no doubt Woz did most of the heavy lifting.

      And Jobs kept most of the money ($4,750), gave Woz $350 and went off to India to achieve enlightenment.

    • Note that creating an arcade game at that time wasn't writing a computer program, it was designing a finite state machine out of discrete logic chips.

  • NeXTstep was all about revolutionizing the programming experience. For the time, it was dramatically more powerful and far easier than the completion. We are basically still using that app environment today ( on macOS and iOS ).

  • The teams he assembled and worked with had a crazy impact on programmers, software development, computer platforms and end users. Even when he way not a developer himself, he looked for very bright people in software development to make it possible to develop new computer platforms. Home computers, early graphical workstations, desktop publishing, UNIX workstation for students, next generation operating systems for personal computers, ...

  • Elon Musk doesn't build rockets either - but it's fair to associate him with SpaceX progress. We rightly blame the top guy when stuff goes pear shaped, and we should be willing to let them get some credit when it goes well.

This is very narrow history. Basically a history that excludes everything that isn't Jobs walk to glory and perfection and domination. Ignore many important points, problems, accidents, alternatives and so on.

NeXT used 'Display Postscript' a display server that was basically a inferior copy of Sun's NeWS system. This was later changed because NeXT was to small and Adobe didn't want to support Display Postscript anymore. Sun of course killed NeWS because they wanted to be a 'standard'. Next didn't care about standards. They had less applications then CDE Unix, and far lower deployment in the 90s.

Objective C is one of many language that you could use to build UI libraries on top of some display system. Objective C wasn't the best or inherently better then many others. Objective C adoption by Next was kind of a historical accident based on office location.

Having something VM based for UI development isn't actually that much of an issue, when the hardware manufacture delivers the OS with the VM included. And usually it his the hardware manufacture that delivers the OS. And VM bases system can be integrated well with the core OS, object oriented or not. And that VM are inherently to slow is also questionable, specially for UI apps that can use C libraries and the Display Server for the most performance relevant stuff.

Apple itself had a very nice system for UI development on Dylan that was arguable better in many way then the Next system. But when Steve Jobs came and they had Next, that wasn't developed anymore.

What Jobs showed of in the late 90s wasn't exactly revolutionary stuff. But Jobs always presents everything as revolutionary.

IPhone development in 2010 working the same as Next development in 1990 is a sign of 'failure', not of success.

  • What current development environment and language would you like to put forward as a successor to/alternative of NeXTstep?

    Some notable applications done in it:

    - Altsys Virtuoso --- this became Macromedia Freehand

    - Lotus Improv --- cloned to become Quantrix Financial Modeler

    - Doom notably was developed on NeXTstep and the WAD/level editor was never ported (though alternatives were later developed)

    - Glenn Reid's PasteUp and TouchType --- two of the nicest apps for working with documents and type I ever had occasion to use

    and a number of ports were done quite quickly and were arguably the best ever versions --- WordPerfect for NeXTstep was done in six weeks and FrameMaker on the NeXT stands out for an interface which I missed when using it on Mac or Windows.

    In particular, I'd really like to have a successor to/replacement for Touchtype.app (and if that could be grown into a replacement for Macromedia Freehand, that would be great) --- what tool would allow developing this easily, and making press-ready files which perfectly match what is seen on-screen?

    • A further example, TeXview.app was a very nice interface for editing TeX documents on NeXTstep, and has since been succeeded by TeXshop.app on Mac OS X and arguably even surpassed (it has many more features, and I can't recall any which are missing).

      Graphical/integrated .tex editing environments are a frequently encountered application type, and there seems to be at least one for pretty much every programming environment --- I'm not aware of a graphical which has feature parity with TeXview.app and which is similarly consistent with its interface (it did win an Apple Design Award back in the day) --- if that's the case, then how can it be that developing with InterfaceBuilder and Objective-C and the NS-objects was so bad? If that's not the case, what development platform would you like to put forward which should have an even better TeX environment?

  • I actually used the Apple Dylan environment. IIRC, it required 48MB RAM -- yes, M -- to launch, which was extremely an extremely large amount at the time -- as in, you probably needed to buy more RAM to run it -- and was itself written in Macintosh Common Lisp instead of self-hosting. UI development? Thing could barely run on a contemporary Mac. Every efficiency claim that was made for it was unproven. Maybe it could have been great, but it certainly needed at least a couple of years more bake time. The NeXT runtime existed and its performance characteristics were understood.

    Virtual memory was a huge issue for UI development with a retained-mode system like DPS; thrashing window contents is very not fun if you want a responsive UI. Apple spent years optimizing VM for this purpose after the NeXT purchase.

    • > UI development?

      The interface builder was written in Dylan and running inside the Dylan runtime.

      The early use of Dylan (actually its precursor language Ralph) was to develop software on a Mac for an external device, like tablets and handheld computers. The development system was an external mainboard attached to a Mac. Apple released eventually a first device with ARM 610 20Mhz CPU, 640Kb RAM and 4MB ROM as a product. A few of the early internal development environments were written in Lisp on the Mac, using a lot of memory (precious and expensive megabytes!).

      > Apple Dylan environment

      That was a prototype of the Apple Dylan environment, later released as a technical preview. It was also never a product. Apple at that time often developed software prototypes and the released product was then recoded as a more efficient tool. The technical preview was for the Mac and for develop software for the Mac.

  • Mac OS development introduced me to Objective-C. Quirky, but I grew to like it. I think it was the first time I 'got' object-oriented programming. I had done Windows programming before and while it was 'fine', writing Mac OS GUI code was so nice. Probably a credit to the frameworks as much as the language, though.

    What other choice did NeXT have?

    C++ was invented about the time NeXT started. Microsoft didn't even release MFC until 1992.

    C...yikes.

    Pascal. Would Steve had went along with that?

    • >Pascal. Would Steve had went along with that?

      I don't see why he couldn't have. It's a far more elegant language and it and other Wirth languages like Oberon have inspired several current programming languages. The standard programming language of the Lisa and Mac was Pascal, and Apple even added object oriented features to it as Clascal (earlier than Wirth's Oberon and not compatible with it).

      https://en.wikipedia.org/wiki/Clascal

    • I did a bunch of iOS development for about 4 years. I found Objective-C way more productive than C++.

    • I'm not saying taking Objective-C was wrong.

      There were many alternatives.

      Pascal is language family. Xerox did Pascal like Cedar at around the same time. Thing Modula-2 existed. Next could have done something along those lines. I'm sure there were commercial versions of that kind of stuff floating around.

      Smalltalk was a big family at the time. Lisp OO system were already common. There were lots of commercial version of that they could have licensed.

      Sun with NeWS had a very extended PostScript that basically allowed you to write most of the application in it.

      There was a lot of stuff going on back then already. History always hides how much stuff was actually happening. Sadly most of it isn't open source, so tons of great stuff goes into a historical blackhole.

      My point is, with the amount of money Next was able to raise and invest in these systems, they could have gone a number of different ways. They made a reasonable discussion at the time.

  • If you're going to bring Sun and NeWS into the conversation, you ought to also mention how after NeWS failed, Sun tried adopting OpenStep for Solaris.

    • Sadly Sun kind of went of the deep end did itself no favors on the desktop. There is a reason they turned primarily into a server vendor. You could also run Xerox Cedar on Sun machines at the time.

      1 reply →

  • I think it's stretching the truth a little to say that Apple "had a very nice system for UI development on Dylan." The Dylan Eulogy site itself claims they hadn't even finished their Interface Builder [1] by the time it was canceled, which NeXT already had at the time.

    As a side note, Apple Dylan seems incredibly limiting and opinionated for no reason. I find it interesting the website lamenting its death only shows screenshots [2][3][4] of the "advanced" editor, rather than any successful applications made using Dylan.

    Also, how can I incrementally migrate Dylan into existing codebases, which were most likely C, at the time? Does it have an FFI? Also, most software engineers mental models at the time were imperative, and they expected them to learn the intricacies of functional programming and object-oriented at the same time?

    That's not even to mention the logistics about running it:

    > Keeping your source code in a fully version-tracked, object-oriented database with full metadata is an idea whose time has long since arrived, yet most of us still spend our time editing text files. Apple Dylan could export and import code from text files, but once you’ve used an IDE that takes advantage of fine-grained object storage of your source, you’ll never want to go back. [5]

    Does this mean I'd need to shape my entire source control system around Apple Dylan? How would diffs and collaboration work? To use my source control system would I have to "export" it to plaintext every time? Text-based AppKit and Obj-C fit right into existing source control systems of the day.

    Also Obj-C and AppKit was already dogfooded heavily within NeXT. The system UI was built using it, as well as all the apps that shipped with the system. You can't say that about Dylan and MacOS 9.

    [1] https://opendylan.org/history/apple-dylan/screenshots/misc.h...

    [2] https://opendylan.org/history/apple-dylan/screenshots/browse...

    [3] https://opendylan.org/history/apple-dylan/screenshots/dynami...

    [4] https://opendylan.org/history/apple-dylan/screenshots/index....

    [5] https://opendylan.org/history/apple-dylan/eulogy.html

    • The platform which Dylan was originally designed for, the Newton had no C to begin with.

      There were two teams fighting for delivering the OS, one using Dylan, other using C++, eventually the C++ team won the internal politics, even though the Dylan one was relatively ahead.

      Newton was programmed in NewtonScript, prototype based OOP.

      Eventually the SDK allowed for C++ native extensions, and on the later version of the OS, before Newton was killed, there was already a basic JIT in place.

      7 replies →

    • I guess it would have been better to say, Apple was working on a nice system for Dylan.

      NextStep had seen more development and existed earlier and was used more, that much is clear.

      I am not saying they made the wrong discussion sticking with it at Post-Jobs Apple.

      My broader point was just that the article leaves out a lot of other stuff happening at the same time.

  • No offense but this comes off as an equally narrow accounting of the historical record.

    • I didn't publish a long article claiming to go into the 'deep history', I just added a few points. I initially had a longer comment but it go unwieldy.

  • Just imagine if NeXT had been based on the Acorn Archimedes platform, instead of M68k.

    There would be no Java on Android, for sure.

    • As someone that knows Java since it exists, I fail to see the connection.

      Additionally, as someone that ported a visualisation framework from NeXTSTEP to Windows, the Acorn would never be able to run NeXTSTEP.

      NeXTSTEP was ridiculously expensive for a reason, in terms of hardware capabilities for the time.

      1 reply →

So, this is why the abomination that is Obj-C is/was used for iPhone/Mac apps. I can't overstate how much I hate Obj-C. I'm so sooo happy Swift has pretty much entirely taken over.

Side note... I feel similarly about the Java to Kotlin transition. Sooo much better. Although, I don't hate Java NEARLY as much as Obj-C.

  • I'm back in Objective-C now for a large mixed codebase for a client and enjoying it. It was my first compiled language starting in 2002 so I've got about equal time with only it as I do with Swift alongside. And I'm spending more time these days in other projects (microcontrollers, mostly) using C so it's a breath of fresh air. I'm finding it generally faster than Swift in Xcode, which is a bonus.

    • I too am back in Objective-C and being away from it for a long while and I'm also enjoying it. For the most part, I really like Objective-C. I think it's a very pragmatic language. My current project is macOS, but the vast majority of my prior work with Xcode and ObjC was with iOS development.

  • To each their own. I'm convinced it's just a visceral reaction to the square bracket syntax. Obj-C remains my favorite language of all time (although I haven't written it in years). Having a high level language that allows you to seamlessly drop into C felt like magic.

    • I'm using Obj-C for the first time in years for a side project. I'm working on making a Wayland backend for cocotron on Linux.

      It is magical the way I can just use the C libraries with all the dynamic goodness Obj-C.

      3 replies →

    • Someone really ought to do an alternative "skin" for Objective-C. That is, define some reversible transformation that can be applied to any Objective-C program that only changes the surface-level details but otherwise results in mapping the exact same program into the later stages of the compiler (and the section of programmers' mental pipeline that comes after the part where their eyeballs and sense of taste are concerned). It's important that the reversible part be adhered to.

      It's really kind of a bummer that we don't have well-known examples of "skinnable" programming languages already. The closest we get are the ones that let you choose your set of keywords for different locales in case you don't want to work with an English-based PL and those block based languages that let you manipulate them as blocks or edit the textual form. I firmly believe that PL skinning would really benefit so many otherwise worthwhile languages that get short-shrift because of negative reactions to the surface-level details. I'm referring in particular languages like those in the Pascal family. (You could also conceive of e.g. a skin for Golang that doesn't have leading caps be the determiner for whether a function is public or not. There are lots of applications for this, and the variations are a non-issue when you have an ecosystem with strong norms about applying a code formatter (like gofmt), which can dictate the format of the canonical on-disk representation.)

      4 replies →

    • For personal use, the only issue I take with Obj-C is how it's considerably less "batteries included" relative to Swift.

      For collaborative projects on the other hand, it can become a handful to maintain if everybody contributing isn't staying on top of hygiene with null checks, type checks, etc. Swift helps to keep all of that under control.

    • Interesting, I guess that part was missed on me since I only really ever used it for iPhone apps and never really had a need to use C directly.

      Also, you're 100% right. The square brackets are what immediately repulsed me and continued to befuddle me even after years of experience with it. Also, everything just feels "backwards" to me if that makes any sense. Coming from Java/C#/JavaScript everything just seemed unintuitive to me at all times. Also, I think this was heavily compounded by using xCode which (at the time) was incredibly laggy. So, I'd mess up the Obj-C syntax and the IDE wouldn't tell me for what felt like forever. Often I'd make a change and hit "play" before the syntax highlighting caught up and that always felt infuriating.

      I last used xCode about 4 years ago and it was still an issue then (even with swift).

      29 replies →

  • Metal is implemented in a mix of Objective-C (CPU) and C++ (Metal Shaders), Swift only has bindings.

    Maybe one it they will rewrite the Objective-C, who knows.

    Kotlin transition is a Google thing, outside Android barely anyone notices that it exists.

  • Did you start with Swift or Objective-C?

    I started with Objective-C and loved it but I imagine that wouldn’t have been the case if I started with Swift.

    • Same. I realize that Objective-C is dated by contemporary standards. But when I first learned it 20-ish years ago it felt like a superpower. I bet it was even more impressive when it first appeared another 20 years further back.

      I think that one of the tragedies of older programming languages is that they survive long enough for people to forget the technical - and technological - constraints that influenced their design. Swift is great, but there are reasons why there weren't any Swift-like languages being developed in or around 1984.

      Similar feelings about Java. It is definitely not my favorite programming language. (I suppose Kotlin isn't either, but it's in the top n.) But it's hard for me to actually hate it. Java walked so that Kotlin could run.

    • I started the iOS development with Objective-C and hated the language. gp's comment is pretty much my feeling as well. I wonder if I'd enjoy Swift so much if it wasn't in contrast to how utterly horrible Objective-C was. These days I still need to revisit legacy Obj-C codebases occasionally or make Obj-C wrappers to interface with C++ code and every time it feels like diving into a cesspit.

    • I was first introduced to Objective-C in 2010 when I worked on my first iPhone app. I've come back to it a half dozen times over the past 15 years since then and always hate it. I only discovered swift a few years ago (when I had to do some iPhone stuff) and was so happy to not need to use Obj-C

  • I did a few things in Obj-C and I only regretted the memory handling stuff. I really liked the performance, reflexibility capabilities (reverse engineering classes is almost trivial) and smalltalk like style. I will not choose it for any project but it gives me the taste that there were more options around C/C++

    • There has been plenty of options back in the day, it was guaranteed that C and C++ would take over everything as they did during the late 1990's.

      Even the Object Pascal to C++ transition at Apple was mostly caused by increased UNIX adoption, and Apple programmers wanting to offer similar experiences, thus MPW was born.

      Same on other platforms, until C and C++ eventually won everywhere, at least until several new kids decided to challenge them in various fronts.

      One that you seldom see nowadays, outside games, is pure GUI frameworks using C or C++.