> temp > à-trier > the-transputer-a-parallel-future-retrobytes

The Transputer: A parallel future

RetroBytes - 2023-11-25

In the 80s one British firm was working of the future of high performance computing, where not 1 processor would work on a task but many. That company was inmos and the processor was the Transputer.

A big thank you to Theo Parmakis (http://parmakis.co.uk)

This video is sponsored by PCBWay (https://www.pcbway.com)

00:00 - Introduction

@IvanToshkov - 2023-11-25

There was a joke at the time, that when Intel created their new processor, they tried to follow their naming scheme and added 100 to 486 but got 585.99382. That's why they named it Pentium :D

@mk72v2oq - 2023-11-25

- How many Intel engineers does it take to change a lightbulb?
- 1.99382, that's enough precision for the answer.

@Joshinken - 2023-11-25

If you’re curious the actual reason intel named it the pentium was because amd and cyrix copied their names because you can’t trademark a number, so intel have their processors and actual name

@surters - 2023-11-25

@@JoshinkenDon't ruin a good joke by naming facts!

@Clancydaenlightened - 2023-11-25

Why did the Intel engineer end up un jail?


For running illegal instructions and possession of illegal opcodes

Lulz

@Carstuff111 - 2023-11-25

I mean, as far as government contracts would be concerned, that is close enough for military grade.....in fact, its too precise, and Intel could have charged more per chip........ :P

@supercompooper - 2023-11-25

I wrote a real time image processing system on these in grade 12. I used SCSI bus on a sun workstation and 8 transputer chips to analyze video from 7 cameras in real time which captured infrared images from metal cooling from steel mills to detect defects.

@wizard-pirate - 2023-11-25

In grade 12? Goodness. What are you doing now?

@supercompooper - 2023-11-25

@@wizard-pirate I've been CTO of a few businesses. Now I own a few and invest in a few. I started coding when I was maybe 11 years old and still code and very passionate re: computer language development. Went to uni for computing/computer electronics and have pretty good skills there too.

@CD3WD-Project - 2023-11-25

​@@supercompoopercan I have a job lol ?

@oidpolar6302 - 2023-11-26

What was the camera resolution one? Was the ADC involved with the analog camera input or it was SCSI/iEEE1394 connected ones?

@supercompooper - 2023-11-26

@@oidpolar6302 each camera was a SCSI target device and another team member handled camera<->SCSI. I implemented SCSI read buffer into the inmos where we had dual ported ram as SCSI data-in buffer. I don't recall exact resolution I think it was 1024 across. It was more complicated as the camera<->SCSI did a convolution of several scan lines and some compression. We were not xfering full frames. More like a convolution of some blocks and cameras were treated more like a fax machine or flat bed scanner. Cameras were IR. As metal cooled if it had an imperfection it would cool at different rates. The system was set up to detect imperfections and characterize them based on their size and cooling rate from time sampling across multiple cameras.

Also the cameras would write to a pseudo device and timestamp their writes so I could asynchronously read from my last ring buffer pointer and I could re assemble the frames of several cameras that way. Was a fun project!

@MarkWebbJohnson - 2023-11-25

In the 80’s I worked with a friend to develop a native 80x86 implementation of occam (called pc-occam). It is a brilliantly simple but powerful language, and so easy to use to write bug-free code. The idea was to make Occam available to students and others who couldn’t afford / get access to a transputer system.

The three things that impressed me most were:

1) Parallel tasks could run on one transputer or another - but the means of inter-task communication (channels) was identical.

2) Multi-tasking was built into the transputer hardware based on task switching on the jump instruction. Simple and efficient.

3) SEQ / PAR. So simple and easy to use.

@cal2127 - 2023-11-28

wonder how well itd run on a modern ryzen system

@paulie-g - 2023-12-03

@@cal2127 Not well. Remember that this was a time when memory access latency was much more in line with CPU performance. To get decent performance you'd have to limit memory use to L1 cache only (and there's no way to do so explicitly on x86/x86-64) and then you'd swamp IPI/IPC (which you also can't use explicitly). Modern architectures are simply not designed for Occam's execution model, which isn't particularly efficient to begin with (it 'won', to the extent that it was used, by being a) tractable and b) giving you a lot of compute by the standards of the day with the transputer). OTOH, you could write an emulator, run the whole thing on a single core and get oodles more performance than it had back in the day (such a thing exists as 'the transterpreter' - a backend for an occam compiler, both now sadly abandoned). Incidentally, there are a few similar projects alive today - a wafer-scale supercomputing chip with tightly coupled processor+memory tiles and a switching fabric, and an ARM-based 'neural computing' set-up in Manchester iirc that is also processor+small memory on a single board interconnected with some sort of ARM bus (not aiming at high performance, just an experiment in mimicking human brain style computation).

@drlloyd11 - 2023-12-27

I used Occam commercially in the 90s. A total dead end of a language.

@paulie-g - 2023-12-27

@@drlloyd11 In what sense? Popularity in industry? Sure. Syntax seems.. inelegant, particularly to people acculturated to Algol-style syntax (Wirth-inspired syntax languages generally suffer from this due to not having a single outstandingly successful language, cf Lua and FreePascal). Conceptually and technically? Very interesting. Unfortunately, the type of problem it is most suited for also typically demands performance and scalability and Occam as-is doesn't deliver on that front, so even if it were updated to have modern tooling, people would likely use more performant languages and kludge CSP constructs on top if they really wanted them. One can see microservices in that way if one squints. Golang is the unobvious and indirect successor.

@drlloyd11 - 2023-12-27

@@paulie-g In the sense it's performance was mediocre compared with using the other hardware available at the same time.
It wasn't useful.

@lrochfort - 2023-11-25

In the early 2000s I worked for a division of DeBeers designing and building diamond sorting machines.

One of the projects was to replace a 1980s optical sorting machine that ran on Transputers. I got to program the Transputers as part of a life extension program whilst we developed the new system. The replacement intel solution had more horsepower, but the Transputer solution was much more flexible and malleable.

More recently, I repaired an ICG photographic drum scanner. The custom electronics in that used a bunch of Transputers on an ISA card for processing the image signals.

I sold some spares to someone in Australia who had a whole cluster running

@stalinvlad - 2023-11-27

USSR Diamonds USSR Oil Wave flag Putin....keep waving flag

@RayBellis - 2023-11-27

Tony Hoare was head of the Oxford University Programming Research Group (in effect the CS department) when I did my Engineering and CS degree there in the late '80s. Part of the CS course was based on his "Communicating Sequential Processes" (CSP) formal language, upon which Occam was built. In later undergrad projects I got to play with quite a few transputers, including ones with audio DACs and 24-bit frame buffers attached. CSP and Occam's communications channels also heavily influenced Google's "go" language.

@valdisxp1 - 2023-11-25

This thing has had some spiritual successors. Basically multicore CPUs with on-chip networking and message passing. Tilera made up to 72 core CPUs. Tilera that go bought by ezchip, then ezchip -> mellanox -> nvidia. And there is also parallella with their 16 core Epiphany chips. Tried writing something for the parallella board back in ~2010s, but was a pretty green computer science student. Could not have imagined that there was stuff like it already in the 80s. Cool video!

@AnonyDave - 2023-11-25

There was also XMOS. Haven't heard much from them for a while. A short bit of time there they had a cheap dev board for their chips but then suddenly they pivoted their designs to weird audiophoolery stuff and I'm not even sure you can get the software to develop for them anymore 🙄

@semicuriosity257 - 2023-11-25

SUN's GX quad polygon 3D system has influenced NVIDIA's NV1.

@garyclouse4164 - 2023-11-25

There's a US company call Parallax that developed an 8 core 32 bit risc microcontroller - the P8X32A Propeller - with each of the 8 cores having 2k ram and shared access to common memory and io through a multiplexor
Parallax codevelopef the chip along with a language called Spin

@0LoneTech - 2023-11-25

​@@garyclouse4164The Propeller is a rather different beast, though. Transputers were designed to run Occam as directly as possible, while Spin is just yet another interpreted language - with the bytecode interpreter residing in ROM on the Propeller. Parallax were known for BASIC Stamp, another interpreter on a microcontroller. Propeller also has no off-chip communications, common to Transputer, XMOS and Epiphany. And while they optimized for handling events asynchronously with sleeping threads and event dispatch, Parallax instead went "interrupts sound complicated, let's not - people can poll now they have 8 processors".

@12Mantis - 2023-11-28

@@AnonyDave I think I have one of those boards. Saw an article on Hack-A-Day about them giving away boards to people with interesting project ideas, now I didn't have a project in mind for it but it sounded interesting so I sent an email saying as much and some time later I got one in the mail.

@KryyssTV - 2023-11-29

It's interesting how computing ended up needing to go back to transputer-like arcitecture to overcome the limitations of the CPU by implementing multiple cores and multi-threading. The guys who designed the transputer really were looking at the big picture.

@PCPSolutions - 2023-11-25

This is a really well timed video. I recently came into contact with a Transputer at the Large Scale Systems Museum in Pittsburgh, PA. They acquired one over the last summer and have been getting it going. What an amazing technology for the time. This was a very informative and well done video, your layout and production quality is top notch!

@EdNutting - 2023-11-28

They’ve got one going again? Nice! Is there a webpage about it or a news article? I’d love to forward it to David May - he’d be so pleased to hear a museum is keeping this stuff alive. He wanted to set up a computing museum in Bristol (UK) but hasn’t quite got there yet.

@hyretechnologies - 2023-12-05

It would be cool to see if there’s a wire-wrapped Warp or CMU Nectar system there.

@Anatoliys_Adventure - 2024-02-18

I didnt know Pittsburgh had a museum like that! I'll have to check it out!

@PCPSolutions - 2024-02-19

@@EdNutting Sorry I had no idea about this comment! No news article but you can check out the Large Scale Systems Museum in NewKensington PA, we have one and Dave has it working!

@PCPSolutions - 2024-02-19

@@Anatoliys_Adventure Check it out! Large Scale Systems Museum on New Kensington. A wide range of really cool old hardware.

@philh9421 - 2023-11-26

My masters degree included quite an in-depth study of the Transputer, and Occam, and it was delivered by Andy Burns at Bradford University, who was on the Occam language panel. A fantastic design concept that in tradition was flogged to the Americans for a song.

@goranaxelsson1409 - 2023-11-25

There actually was another OS for transputers, Trillium.
I once had the source code for it but a disk crash 30 years ago made me loose it.

From the beginning of an article...
"Trillium is an operating system designed for a
loosely coupled, message passing ensemble architccture,
such as the hypercube class of machines. It has
been implemented for the FF’S T-Series[l], a transputer
[2] based machine. Trillium presents a uniform
programming environment that extends out of
the parallel processors to encompass the front end
computer and and personal workstations accessible
over a network. C and Fortran compilers are available
along with a suite of backend development tools
including a transputer board simulator."

A couple of friends and I actually had it running on our homebuilt transputer hardware back in 1991.

@majorpentatonic2310 - 2023-12-01

There was also Minix. I had it running on my transputer system a few years ago.

@llothar68 - 2023-12-24

Homebuild Transputer, maybe the only thing that i was interested more in 1987 then losing my virginity.
But unfortunately, all tries with my own first designed and build Z80 system showed me i will never be able to this fine tuned skills of soldering.

@Error6503 - 2023-11-26

I worked at the Inmos Newport fab during the Thorn EMI years; I wrote test routines so if you have a duff chip it might be my fault 🙂 The transputer had very few opcodes (they were fundamentally 4 bits long) but it wasn't a RISC chip to anyone except marketing; Its instruction set natively implemented multi-threading, which was very complex, in order to allow single chip parallel algorithms. In my time there we went from successfully introducing 20Mhz T800s to failing to introduce 30Mhz T800s, with some cool new stuff cancelled along the way. Meanwhile Intel had gone from 12Mhz 386s to 50Mhz 486s in the same period, which really showed what we were up against.

@12Mantis - 2023-11-28

So what sort of cool stuff got cancelled?

@Error6503 - 2023-11-28

@@12Mantis The most interesting was a vector processor in place of the FPU on the T800; it could have been a GPU before the world had ever though of a GPU. There were also plans for other dedicated interface processors alongside the M212 (which had a disk interface on chip). And the project which eventually produced the T9000 was restarted from scratch at least twice.

@JamesOversteer - 2023-12-14

@@Error6503I grew up near and spent a lot of time completely unaware of that spot in Newport. Interesting to hear some inside knowledge, thank you for sharing.

@markbooth3066 - 2023-12-16

I always thought it was a shame that T9000 link packets weren't just a little bit longer. If they had been, the technology would have been shoe in to implement ATM. The C004 switch chips were ahead of their time, and the C104 was the logical extension of that, but without being able to tap into that burgeoning telecoms market, DS-Links never got the economics of scale to become a world beater.

@abarratt8869 - 2024-03-24

Thinking they could stay in the fab-ownership game was what finished them. Intel were pouring $billions into silicon process development, Inmos had £0.50... It was never going to work.

@cohort6159 - 2023-11-25

I built a 64 processor reconfigurable transputer system in grad school. I went to the Transputing 91 conference where the T9000 was announced. I never fooled with them after I graduated.

@larsp.sorensen4035 - 2025-08-06

I actually had a T9000 in my hands back then, about 1994. It was a prototype. The virtual links never work. It also had a tendency to latch up now and then. Funny enough, after cool down it work again. I never heard of any production T9000.

@rtechlab6254 - 2023-11-25

Used a crapton of these in my first job. Crash test systems, predictive failure analysis and signal processing. We went through thousands of them. Development used two Meiko dev systems and one very much like if not the same type unit you have. We used the ISA cards and TRAM modules as well as our own cards. Youd have a system with a couple of T440s called PMMs then boards with high end Analog Devices ADCs and filters and a single T400 with a chunk of local RAM. The comms and booting was handled on the power station units, by a 486DLC, pc104 based machine. Most power stations in the UK ran these.

@rtechlab6254 - 2023-11-25

You might get some luck reaching out to Prosig in Fareham UK

@stephenhookings1985 - 2024-02-26

I did something similar (designed hardware to memory mapping into two transputers on the Meiko image bus with some A110 (I think) digital convolution chips and a ton of Intel i860s (whilst waiting for the T9000 who mysteriously turned up in a Terminator film!).

I shrank my system down to run on an M10 in the back of a Armoured Personnel Carrier and observe IR images then identify potential military vehicles. I really like what Meiko did - I am sure I worked with David May (he uses to call me Derek of Derek and Clive fame) and Neil Coulson - both exceptional Meiko engineers. Meiko is not a canteen automation system - look at end of some belt driven tray conveyor belts. I did get my hands on an early Atari system, as well as the TDS boards for the 80286's. But the Meiko with its programmable MRUN took things to the next level. A few years later one DEC Alpha (ruggedised Military spec) could so same job as 6 i860s with 2x transputers for comms, the image capture cards and the Sparc 2 on the VME acting as display.

Progress

@andrewstones2921 - 2023-11-25

Wow this brings back some memories. I used to work for a company in Portsmouth in 1989 that had just moved from Bristol and was manufacturing and testing these transputer boards, we were told that they were used in high end simulators. I didn’t last long, I worked as a test engineer and shortly after joining the company I made some mistake on a batch and they shipped with some problem and I was rightly blamed and fired.

@stephenhookings1985 - 2024-02-26

FPS America filed for Chapter 11 after their Array Processor mk 2 failed - they blamed it on the Transputer. So don't feel so bad.

@ardentdrops - 2023-11-27

I can't believe the TIS-100 actually exists! I thought it was an original idea 🤯

@CaptainSlowbeard - 2023-12-20

Was just about to say the same thing!

@enthusiasticgeek7237 - 2024-02-02

me too!

@halfsourlizard9319 - 2025-02-01

Nothing makes me happier than hearing 'full formal methods' -- things like CPUs and compilers and garbage collectors cannot afford to have bugs.

@honkhonk8009 - 2025-06-24

Intel hears this and decides to have a bug in their AI branch prediction.
Clearly there isnt enough DEI hires

@halfsourlizard9319 - 2025-06-24

​@@honkhonk8009Um didn't come here for political bullshit.

@caryj57 - 2023-11-25

The strangest thing about working with transputers was the memory address range. It went from -maxint .. +maxint so 0 is a valid memory address right in the middle! You had to redefine what NUL was because dereferencing *0 was perfectly valid.

@renakunisaki - 2023-11-27

Nice, an addressing system even weirder than x86 segmented mode.

@rosomak8244 - 2024-02-26

@@renakunisaki Look up the Zilog Z8000. You will be pleased. Another gem in this regard is the Power architecture. However there you can safely ignore the oddities.

@jfbeam - 2024-11-07

Good point, but zero is a valid memory (bus) address on PC's, too. OSes protect it - as part of page-zero - so NULL references can be detected. I've used "zero" on loads of simple devices. (and some not so simple... moto 68k)

@StanleytheCat-v8z - 2025-01-23

Oh yeah, read that in the ST20 ISA pdf! Whoever came up with THAT idea must've been on something....

@5urg3x - 2023-11-29

This kind of stuff should really be preserved and enter the public domain after a certain amount of time; the development boards and software and all that, the source for everything as well. I think the same should apply to older video game consoles and other computing devices, etc.

@kenfuller9907 - 2023-11-25

Very nostalgic for me. I did my masters degree in Parallel Computing Systems in Bristol at what was once called The Bristol Transputer Unit. The last year the course was run. So I’ve experienced the flashing curser indicating deadlock first hand…..a lot!

@RetroBytesUK - 2023-11-25

Was David May one of your lectures?

@kenfuller9907 - 2023-11-25

@@RetroBytesUK not that I recall but the transputer was pretty much dead by that time and there was more emphasis on parallel paradigms like neural networks rather than hardware.

@richard_loosemore - 2023-11-26

Was that at Bristol University, or the Polytechnic? I worked in the Bristol Transputer Centre, located in the Poly, during the peak period of 1987-1990. We had tons of transputers, but couldn’t use most of them on accounted of lack of dev tools.

@richard_loosemore - 2023-11-26

I also dated Ian Barron’s daughter Clare for a while. Very sweet girl, but I was kinda useless as a boyfriend. 😀

@kenfuller9907 - 2023-11-26

@@richard_loosemore it was called the University of the West of England when I went there but yes it was the Poly.

@ebrombaugh - 2023-11-25

Back in the 1990s I worked for a small R&D company that was developing a radio analysis system based on transputers connected to a huge rack full of Xilinx 3000-series FPGAs. We had a couple PhDs writing code for the system in various flavors of parallel languages (Fortran & C IIRC) to do signal analysis. Not sure if it ever went anywhere, but it sure generated a lot of interesting hardware.

@Aim54Delta - 2023-11-26

I know some people who were working on some signal capture stuff that went toward the MALD and similar programs.
While probably not the same company, I am sure some of the same people, or students of those people, were working on similar stuff.
I think FPGAs generally outshined the transputer in time - although... the simplicity of the transputer would make it scale into high frequencies rather easily and it would not be difficult to commission a custom series on new technologies of die lithography.

A bit much for a dod contract, but depending on the gusto of the groups involved and whose signatures they could get - not outside the realm of possibility.

Generally, though, I would say that modern FPGAs are the successor to the transputer in many ways. If not literally, then functionally.

@ebaystars - 2023-11-30

satellite downlink interception/microwave point to point/ 🙂

@piotrd.4850 - 2023-12-02

@@Aim54Delta think that every decade or so, some projects and concepts should be revisited with updated tech, manufacturing and other advancements.

@TawaSkies - 2023-11-25

The Atari Transputer was one of my dream machines in the day. I remember reading an article about one of the UK's universities building a machine based on these with no ram. A bit fuzzy on details time always warps memories lol, might have been Cambridge anyway loved your story on theses, nice bit of research and history. Looking forward to you being able to get your machine running.

@dipaoloboat - 2023-12-02

Thank you for enlightening me as to what a transputer is, I had heard the term back in the 80's but never knew exactly what it was. Around the same time, maybe a bit earlier about 1984 I was an engineer working for Loral Instrumentation of San Diego California on the development of the Loral Dataflow computer, one of the first parallel processing computers, which was based on the concept of a "data freeway" called the "Tag/Data bus". The Dataflow computer contained up to 512 processing nodes, which like transputers each had own local memory and ran their own code. The Dataflow computer was intended for processing an incoming stream of realtime telemetry data. Each node was programmed to grab from the data bus one or more data elements having predefined "tag" values, do its processing on that data, then put the result back out on the bus with its own predefined tag value to be picked up by either an output device or one or more other processing nodes to perform further processing and put their results back onto the data bus with a unique tag value identifying its result to any other nodes needing that result, etc., etc. Often a processing node needed more than one specific "tag" of data and captured the most recent of its assigned "tagged" data elements, processing a single "result" when all values its algorithm needed had been captured from the data bus. So unlike transputers which were matrixed together the Dataflow nodes were all attached to one large high speed parallel data bus and a parallel tag bus (taking the place of the address bus in a conventional Von Neumann architecture). Its intended uses were for realtime processing of telemetry streams from spacecraft, aircraft, and signals intelligence systems. Those were exciting times, we felt back then like we were inventing the future and having a lot of fun doing it. :)

@GeekBoi - 2023-11-25

I always wondered what the deal was with the transputer cards was. Always saw them in the Amiga magazines but never really saw any articles or documentation on what they were used for. The 286/386 PC cards always made sense to me but these were just a concept I could not follow just based on the advertisements. Thank you for the interesting history. Having worked on parallel computing clusters in my professional life I now have a much better appreciation of what these cards could do all the way back then. Also loved the razor joke 😂

@GodmanchesterGoblin - 2023-11-25

This is a really well set out and comprehensive introduction to the Transputer. Well done! Massive bonus points for the BCPL mention. That was my first encounter with a serious programming language back in 1981, with the compiler and linker generating code for a TMS9900 based system. It's a relatively simple language but well suited to older processors with their more limited memory and instruction sets. As for the Inmos founders, the only name that I recognised was Ian Barron. The Transputer was way ahead of its time. Too bad that I never did any hardware design with it.

Also, interesting footnote... Inmos was also the company that created the IMS-G171, the colour palette chip used in the original IBM VGA implementation and its many clones. There were other widely used variants too such as the G176.

@NotMarkKnopfler - 2023-11-28

There is a chip that is still in production today that is similar. The Green Arrays GA144 has 144 cores on the same chip and they all run their code in parallel. The cores on the edges have access to the chips pins. The native assembly language is Forth. They are stack based cores. 👍

@Pest789 - 2023-11-25

I was heavily into Atari 16 bit machines at the time and I followed the Atari transputer development as closely as possible. From my perspective back then, it seemed a lot more like they never had a product to sell, rather than a product to sell that nobody wanted.

@alexhajnal107 - 2023-11-27

My recollection is that they were always busy developing new things, not so much on releasing them. What little did make it out the door was mostly minor upgrades on existing kit so they ended up falling further and further behind.

@kgblankinship - 2023-12-01

'Transputer' - sounds like a device out of Star Trek.

@honkhonk8009 - 2025-06-24

on foenem grave

@Nerfty-l9k - 2023-11-25

The transputer was definitely ahead of its time. Intel were waging a MHz war for much of the 90s/00s, while the transputer designers foresaw the need to go parallel. Looking at modern CPUs, inmos were right.

Programming languages weren't designed for massively parallel computer systems (many still aren't), and occam introduced new paradigms that would be necessary based on Communicating Sequential Processes (CSP) with channels and guards. AIUI, CSP was developed after Hoare abandoned "monitors" as being unworkable. Monitors found their way into the java language as a primitive. Quite easy to shoot yourself in the foot with them.

Echos of this work lives on in languages like Go, which have support for channels.

@0LoneTech - 2023-11-25

Chapel is another programming language built for distributed computing.

@johnjakson444 - 2023-11-26

For anyone that is a hardware and software engineer, CSP/Occam and hardware description languages like Verilog are on a continuum, anything can be modeled in either, but Verilog is more suited to modelling describing real hardware while Occam more suitable for modelling the parallelism structure, sort of 2 sides of the same coin. See also HandelC.

@petergoodall6258 - 2023-11-27

In the mid-to-late ‘80s worked for Integrated Arts (Tiny Grated Rats) in Australia. We were building a broadcast video animation system, which had hundreds of t800s. Very difficult to write something like an operating system for the transputer. No exception handling. If you wanted to kill a process, you had to write ‘stop’ all over its address space. I wrote a parallel memory management system for the other folks to use. So much fun for so little result :-) Occam - a British intellectual programming language.

@taskell4449 - 2025-06-14

T800 huh what a great choice for a name

@Distinctly.Average - 2023-11-25

I was at Hatfield Polytechnic back in the day. It was astonishing at the time to see a computer generating live Julia sets. This was on their small transputer setup. It was a shame it died off as many thought at the time it was the future.

@RetroBytesUK - 2023-11-25

If it had not been starved of funds at a key time in its development they probably would have been right, and it would still be a major player.

@6581punk - 2023-11-26

Parallel processing was a reaction to the fact that 32-bit processors weren't showing the performance gains expected of them. But that soon ceased to be the case, taking the wind out of the sails of parallel processors. Doing software development across multiple execution cores is vastly more difficult and still is even to this day. So nobody wanted to go down the parallel processing route until they absolutely had to. Real life applications are a lot more complex than generating fractal images.

@Distinctly.Average - 2023-11-26

@@6581punk That depends on the application. Breaking a task down into sub tasks and doing them separately is something we have been doing for a very long time. Having a well designed OS that can deal with that is the key. I’ve worked in the mainframe arena for many years now so have worked on applications that do just that.

@Distinctly.Average - 2023-11-26

@@RetroBytesUK Sadly a lot of things that could have been great have failed not because of the technical issues, but financial or poor management. I recently listened to a talk by some of the original Amiga team and they honestly felt some of their plans were way ahead of their time and would have come to fruition had Comodore not failed. I blame John Carmack for creating a game that the Amiga couldn’t play lol

@andywilkes3333 - 2023-11-27

The company sent a couple of the team to the college in Norwich at the time that the Transputer was big news in the electronics mags. I remember, at a time when a single fractals took minutes to be drawn on the computers available to me that they showed a 'flying view' over a fractal landscape.

@hessex1899 - 2023-11-25

When I was in school I used to pour over the specs of the transputer and daydream about getting access to one. Thank you so much for doing this video. :)

@RetroBytesUK - 2023-11-25

You're welcome.

@mikehibbett3301 - 2023-11-25

What a lovely project to work on! I have a transputer somewhere in my component collection, which I was given while working in Cambridge in the 1990s. I hope i find it again :)

@MoreFunMakingIt - 2023-11-24

Such a great video. I learn about things I could only have dreamed existed in my wildest imaginings! Superb.

@kazzle101 - 2023-11-25

24:07 I now see why I got to play on an Atari Transputer Workstation for a week at Sheffield TriTec as a technology demonstration event sort of thing... Not that I knew what I was doing or anything about it really, like the basics of how it worked. Very impressive computer though, a big tower box in that Atari Grey colour and a massive Monochrome high resolution CRT monitor.

@chriswareham - 2023-11-25

Wow, I knew some of the history of the transputer but not the details. I love the foresight of designing the programming language and implementing a compiler for it first. There's some great British contributions to computing that are in the shadows, and you mentioned one them in passing - BCPL. This was the inspiration of the B programming language, that evolved into C. I still have the definitive (only?) book about it, written by Richards and Whitby-Strevens.
As for the fate of Inmos, another example of Thatcher's blind belief that the private sector always runs things better than the public sector. Same blindness that ruined so much of Britain's economy, such as the rail industry (not just the railways themselves but the supporting industries like train manufacturing).

@RetroBytesUK - 2023-11-25

The approach of the Thatcher goverment did do a lot of longer term economic damage, and snuffed out areas of development that could have generated whole sectors of the economy. There jusr seamed to be an absense of medium and long term thinking.

@Cybergorf - 2023-12-11

BCPL did earn itself a very bad reputation within the Amiga crowd.
BCPL was a nice academic language and TripOS, the operating system build upon it, saved the early Amiga by providing the missing disk operations and a shell.
But BCPL was stuck on 32bit pointers that did not take into account that the Motorola 68000 (as all other common processors) would address memory in bytes (8bits).
So to get to your real address in memory you had to multiply by 4 - or to get to your "BCPL-pointer" divide by 4. This was a huge burden for programs written in C or Assembler and a constant reason for bugs.
Commodore later spent quite a lot of resources to replace the BCPL code with rewritten C code.

@chriswareham - 2023-12-11

I didn't know that BCPL was used for the early Amiga operating system - that's fascinating! I do recall having to deal with data alignment issues with some C code that was ported from a 16 bit system to a 32 bit one, but thankfully the compilers got better and started to handle the alignment in structures itself.

@Cybergorf - 2023-12-11

@@chriswareham Yes: the original Amiga crew (when it was still a startup before Commodore acquired it) tried to build up their own operating system. They got a preemptive multitasking kernel (Exec), developed by Carl Sassenrath, and the GUI (Intuition) by RJ Mical, but they were lacking proper device handlers and a command line interface. So Commodore contracted British Metacomco to "transplant" BCPL based TripOS onto the Exec kernel.
So the System was from the start a weird mixture of C structures and BCPL structures, which made it overcomplicated for developers ...

@paulfrindle7144 - 2023-11-27

Back in the late 1980s I was a co-director of a company charged to design a large scale digital audio mixing system for a multinational organisation. At the same time close friends of mine were working for another company who were trying to do the same thing. They were wedded to the early transputer system as its core processing engine - they had been sold the concept of it.
However it was pretty clear that politics in the UK would eventually clobber the transputer initiative - the UK was supposed to become a 'trading nation' and it was therefore bound to be sold off and eventually fail etc :-(
So instead, we produced a large scale array of our own home-grown custom processors, specifically designed for the tasks we needed, with several added custom features, which allowed our advanced development system to function, which would never have been viable with the transputer or any other processor at that time. Needless to say we succeeded, and sadly the other team failed.
But the moral of the story is that the volatile political system in the UK from the 1980's no longer allowed any government backed home grown technical development over more than one parliamentary period - simply because no politician during a 4 year period would ever understand its significance - and the mantra of 'trading nation' is now burned into our politics of all flavours. This is what would kill off the transputer and we had predicted it correctly. It's such a pity :-(

@hellbreakfast - 2024-02-29

The thing about the politicians is absolutely wild to me, because developing computers was such a good play towards success. So much great tech we use today, including the internet, is simply because one government or another backed and developed it. I actually only recently found out how on the cutting edge the UK used to be in regards to computer tech- I'm not from there, and of an entirely different generation, so I never would have guessed until taking an interest in vintage computers. Blows my mind.

The closest thing I can think of right now, as an American, is how we've gutted NASA to near non-existence, and handed off space travel to private interests. It's gone... Well, exactly as one may expect, to be honest.

@m-u-l8r747 - 2023-11-26

Back in the 80s there was a front cover of a gaming magazine (I can't remember which one) which featured screen shots of a game for the transputer. It was a multiplayer 3d shooter where each player had a different military vehicle, a plane, a tank, etc. and they had to take each other out. I remember seeing it and being amazed at the graphics (very rudimentary 3d shaded) at the time, as well as the concept of multiplayer 3d shooters with players on different machines - it was a forerunner of online gaming. I had no idea what a transputer was at the time and this is the first time I've heard about it since. Thanks for the great video.

@zh84 - 2023-11-25

I remember advertisements in Byte for Microway transputer ISA boards. They had the Monoputer, with a single transputer on it, the Quadputer, with four, and the Number Smasher, which had a transputer doing I/O for an Intel i860 - a combination of two powerful and now obscure technologies!

@qu1j0t3 - 2025-06-28

Adding to the replies from people who actually used these: In the 1980s I bought two T800's on 1MB TRAM modules, and used them in Mac II hosts to run TeX and METAFONT (I ported them to the Transputer using modified Pascal to C translators and Parallel C) and my own ray tracing code. They were signficantly faster than the Mac host. A very elegant architecture.

@keirthomas-bryant6116 - 2023-11-25

Another amazing video, thanks! I was in the Atari orbit across the 1990s because I had an Atari ST. (Well, several Atari STs... They weren't very reliable.) And so when you mentioned the Atari Transputer, a load of lightbulbs went off in my head. It was featured quite a lot in the Atari ST magazines. There was such a push in the world of Atari for serious computing tasks that belies the memories of the Atari ST as being a weaker version of the Amiga. Lots of us felt the ST was very Mac-like, and certainly much more than the Amiga. Atari already had the music scene sewn-up and there was quite the push into DTP, too, if I recall. Atari platforms by this point were just more capable, and this affected the culture, too. The Amiga was very good at some things, but this caused is community to be very inward-looking. The Atari ST, on the other hand, transformed into more of a platform, in the way that the IBM PC would be. This was because of those MIDI ports, or the high-res monochrome graphics mode and monitor that facilitated pin-sharp DTP. I kinda wish Atari had continued with this philosophy but the history of all the big home computing names in the 1990s was essentially self-destruction, as we all know.

@medes5597 - 2023-11-25

As a graphic designer, I was always impressed that Atari users tended to know their computers absolutely inside out. They knew how to make art programs do anything you wanted. I remember seeing a floppy disk that receated Andy Warhol's Amiga art experiments perfectly on the Atari. If you were a designer on a budget, an Atari ST was a must have tool. I honestly don't understand why Atari didn't make a bigger deal of its art capabilities. The Amiga and Mac absolutely cleaned up with artists and it was down to the users to show that an Atari ST could do more than a Mac and as much as an Amiga for less than half the price.

I've always felt like the knowledge of the atari userbase has been hugely understated. Atari being kind of useless meant their users took ownership of their machines.

The ST was like an amazing little creative toolbox and was cheaper than any of its competitors. And yet I've never seen a single advertisement that targeted that audience.

The ST was honestly one of the best computers I've ever owned for art.

I know it wasn't quite what you were talking about but the amount of missed opportunities around the ST just always make me wonder what could have been.

Even the head of Atari UK said "by the time word had caught on that the ST was an amazing machine for music, being used by fat boy slim and white town to get number one singles, we'd stopped making the bloody things. Only the surplus stores and second hand market got the benefit. Atari had trouble telling people anything about it. "

@another3997 - 2023-11-26

I'm a fan of both companies product, but your assertions about the ST being "more capable" and Amiga "inward looking" are simply not true. The fact that GEM looked like the Mac desktop was why Digital Research were (unfairly) sued by Apple, and it was changed accordingly, except on the ST, because Apple didn't see Atari as a credible threat. And they were correct. Like the early Macs, the ST OS was fairly primitive, single tasking, quite limited and no CLI until Atari adopted MINT. The Atari hardware suffered in comparison to the Amiga's graphics, and outside of MIDI, it's built in sound. Even the later inclusion of a bitter and DSP on some models didn't help, as virtually nothing took advantage of them. And whilst the Amiga got the 68040 CPU, Atari never went beyond the slower 68030. Then the Amiga had the Newtek Video Toaster series with Lightwave 3D, which catapulted it in to TV and video production. They could be linked in to render farms, and a high end MIPS R4400 based parallel render system was developed by Desktop Technologies just for rendering Amiga Lightwave animations. Let's not pretend one system was ultimately superior... they both had their blaze of glory, and both died painful deaths.

@keirthomas-bryant6116 - 2023-11-26

@@another3997 I think you missed my point but maybe I didn't explain properly. When I say the Amiga system is inward looking, what I mean is that its built-in chips for sound and graphics meant that people had all they needed. More than they needed, in fact! Because of this, most people who bought an Amiga in its heyday were in the game for what it could do on its own, and by itself. But because of design decisions with the ST – and I mention its MIDI ports and the Mac-like high-res monochrome SM124 monitor in my comment above – a culture grew up where the ST was more the centrepiece of an external setup. I think this is shown really well by comparing the music scene that arose around the Amiga and the ST. With the Amiga, it was mostly about mod music, because the Amiga had a DSP. An Amiga was all you needed! With the ST, it was about loading up Steinberg 24 and powering your synthesisers. The ST was just one component of a wider system. In other words, the Amiga was still very much a home computer in that it was all about what it alone was capable of. But the ST was indicating the direction of travel for the coming decades, where a computer would drive peripherals – where a computer wasn't inward-looking but outward-looking. This is my point. And a few other design decisions of the ST underlined this approach and culture, too, such as MS-DOS floppy compatibility (the Amiga formatted disks in the Amiga way because why would you ever need anything other than an Amiga?!). Plus there was the GEM desktop, which was much more Mac-like than Workbench. Workbench carved its own path in ways that you never found in any other GUI like Windows and, again, this indicates the inward-looking direction of the Amiga. If you used GEM then you could switch to MacOS and Windows pretty easily. If you used Workbench you'd be wondering why everything in Windows or MacOS was so weird. I realise that, as you say, video capture and animation opened up the Amiga to a certain kind of low-budget TV and movie production workflow. So it did catch-up to the "outward-looking" culture of the ST. But the whole video scene was some years after the launch of the Amiga. Summary: People bought an ST for doing the things they were interested in. People bought an Amiga for the things you could do with an Amiga. That's a subtle but important difference.

@alexhajnal107 - 2023-11-27

@@another3997 The result of the Apple/DRI lawsuit didn't apply to Atari because they (Atari) had done the bulk of the m68k porting and were thus given full development rights without having to license the changes it back to DRI. For legal reasons I won't even pretend to understand this meant that they were not bound by the suit's result. Not sure why Apple didn't sue Atari directly though.

@paul_boddie - 2023-12-08

@@another3997 "and a high end MIPS R4400 based parallel render system was developed by Desktop Technologies"

That would be DeskStation Technology who had already released a number of MIPS-based workstations, initially aiming to ride the Advanced Computing Environment bandwagon, eventually having to settle for making Windows NT workstations based on MIPS and later Alpha. Interestingly, NewTek and DeskStation were both Kansas-based companies.

@assifmirza130 - 2023-11-27

Great video, really well researched and presented. My run-in with the Transputer was at UMIST in 1988 - we had a lab with 20 Atari 520STs as workstations, and access to a number of transputer system. I can't remember how many or how they were hosted because all we undergrads ever got was a bunch of lectures about them.

Thanks for reigniting those lost memories!

@neogen23 - 2023-11-28

Only a year ago I listened to discussions among Forth groups on a parallel chip design that uses the exact same grid style for layout and communication between cores. Who knew it had already been done by others (and failed). Also, reading recently on Haskell parallelization, par and seq and relevant functions are front and center, along with functions that serve exactly like alt. So, thanks for the history lesson. It has been eye opening. Never knew these systems even existed. Keep them coming!

@paulie-g - 2023-12-03

The Forth chips you're talking about, whether they're the Green Arrays (iirc) or Propeller are a different lineage and different purpose - they're essentially programmable realtime IO and run fairly independently of one another with not much processing power available or required. Seems similar, but isn't.

@interlocking-joinery - 2024-01-14

I did a lot of work with the Novix chip in the 1980's which had Forth as it's native instruction set (assembler). Is this the chip you are referring to?

@enitalp - 2023-11-25

I was an official Atari dev, and I have a good collection of Atari hardware, one day I saw 2 ATW800 Atari transputer, and they were for sell for cheap, but at the time I did pass the offer. I still regret it. There was a picture at a time, at Kodak ,with a wall of ATW800 doing images manipulations.

@andygardiner6526 - 2023-11-25

I remember seeing a machine at university - a student's father worked at Inmos and got a development system running in his room for a couple of hours! The classic ray traced Newton's Cradle demo at the time just blew our comp-sci nerdy brains. :-) For a term we studied Occam and I still have a copy of he Occam programming guide which was expensive, very thin and yet managed to have many pages with "this page is intentionally left blank" printed in the middle which made it the worst value book I have ever purchased!

@RetroBytesUK - 2023-11-25

Always annoying to feel you got ripped off by buying the text book they told you to buy.

@zosxavius - 2025-07-01

I've been studying computer history for decades and these videos are SOOOOOOOO GOOD.

@0Zed0 - 2023-11-25

Just started this video and it brings back memories. At the time of the transputer my Dad was working for a smallish print firm that printed some marketing materials for Inmos and knowing I was into computers he brought home some for me. I had a booklet on the ITEM400 and ITEM360 (Inmos Transputer Evaluation Module) 400 or 360 MIPS. Now back to the Video.

@0LoneTech - 2023-11-25

Three modern day cousins to the Transputers are XMOS, Epiphany and GreenArrays. Mailbox registers, as used in e.g. the RP2040 for interprocessor messages, are also related.

@julianskidmore293 - 2023-11-26

I used the UEA's 9 Transputer rack in the late 1980s to do my BSc dissertation: "The Implementation of Neural Nets on a Transputer Rack" which implemented a fully-interconnected, parallel neural net on 20MHz (ie 5MHz x 4) T424s. Amazing system!

@HectorGrey223 - 2024-05-04

The idea of channels has of course been brought into asyncronous programming in modern languages because it's still a really good idea to make sure that separate threads don't get to share variables.

@Vanders456 - 2023-11-25

I very very very briefly worked for Quadrics, the company that was spun out from Meiko to develop the cross-bar switching fabric into a supercomputer interconnect. There was a Meiko CS kicking around the office. I wish I had known more about it at the time: I'm fairly sure I could have got my hands on it if I had asked.

@howardjones543 - 2023-11-27

I always thought the name Computing Surface sounded super bad-ass.