Saturday, April 26, 2025
HomeArtificial IntelligencePhillip Burr, Head of Product at Lumai - Interview Sequence

Phillip Burr, Head of Product at Lumai – Interview Sequence

Phillip Burr is the Head of Product at Lumai, with over 25 years of expertise in world product administration, go-to-market and management roles inside main semiconductor and know-how corporations, and a confirmed observe document of constructing and scaling services.

Lumai is a UK-based deep tech firm creating 3D optical computing processors to speed up synthetic intelligence workloads. By performing matrix-vector multiplications utilizing beams of sunshine in three dimensions, their know-how affords as much as 50x the efficiency and 90% much less energy consumption in comparison with conventional silicon-based accelerators. This makes it notably well-suited for AI inference duties, together with massive language fashions, whereas considerably lowering vitality prices and environmental affect.

What impressed the founding of Lumai, and the way did the thought evolve from College of Oxford analysis right into a business enterprise?

The preliminary spark was ignited when one of many founders of Lumai, Dr. Xianxin Guo, was awarded an 1851 Analysis Fellowship on the College of Oxford. The interviewers understood the potential for optical computing and requested whether or not Xianxin would think about patents and spinning out an organization if his analysis was profitable. This acquired Xianxin’s inventive thoughts firing and when he, alongside considered one of Lumai’s different co-founders Dr. James Spall, had confirmed that utilizing gentle to do the computation on the coronary heart of AI may each dramatically enhance AI efficiency and cut back the vitality, the stage was set. They knew that current silicon-only AI {hardware} was (and nonetheless is) struggling to extend efficiency with out considerably growing energy and value and, therefore, if they might remedy this drawback utilizing optical compute, they might create a product that clients wished. They took this concept to some VCs who backed them to kind Lumai. Lumai lately closed its second spherical of funding, elevating over $10m, and bringing in extra traders who additionally imagine that optical compute can proceed to scale and meet growing AI efficiency demand with out growing energy.

You’ve had a formidable profession throughout Arm, indie Semiconductor, and extra — what drew you to affix Lumai at this stage?

The quick reply is crew and know-how. Lumai has a formidable crew of optical, machine studying and knowledge middle consultants, bringing in expertise from the likes of Meta, Intel, Altera, Maxeler, Seagate and IBM (together with my very own expertise in Arm, indie, Mentor Graphics and Motorola).  I knew {that a} crew of exceptional individuals so centered on fixing the problem of slashing the price of AI inference may do wonderful issues.

I firmly imagine that way forward for AI calls for new, revolutionary breakthroughs in computing. The promise of having the ability to supply 50x the AI compute efficiency in addition to slicing the price of AI inference to 1/tenth in comparison with at the moment’s options was simply too good a chance to overlook.

What have been a number of the early technical or enterprise challenges your founding crew confronted in scaling from a analysis breakthrough to a product-ready firm?

The analysis breakthrough proved that optics might be used for quick and really environment friendly matrix-vector multiplication. Regardless of the technical breakthroughs, the most important problem was convincing folks that Lumai may succeed the place different optical computing startups had failed. We needed to spend time explaining that Lumai’s method was very completely different and that as a substitute of counting on a single 2D chip, we used 3D optics to succeed in the degrees of scale and effectivity. There are after all many steps to get from lab analysis to know-how that may be deployed at scale in a knowledge middle. We acknowledged very early on that the important thing to success was bringing in engineers who’ve expertise in creating merchandise in excessive quantity and in knowledge facilities. The opposite space is software program – it’s important that the usual AI frameworks and fashions can profit from Lumai’s processor, and that we offer the instruments and frameworks to make this as seamless as potential for AI software program engineers.

Lumai’s know-how is alleged to make use of 3D optical matrix-vector multiplication. Are you able to break that down in easy phrases for a basic viewers?

AI programs must do quite a lot of mathematical calculations known as matrix-vector multiplication. These calculations are the engine that powers AI responses. At Lumai, we do that utilizing gentle as a substitute of electrical energy. This is the way it works:

  1. We encode info into beams of sunshine
  2. These gentle beams journey via 3D house
  3. The sunshine interacts with lenses and particular supplies
  4. These interactions full the mathematical operation

Through the use of all three dimensions of house, we are able to course of extra info with every beam of sunshine. This makes our method very environment friendly – lowering the vitality, time and value wanted to run AI programs.

What are the principle benefits of optical computing over conventional silicon-based GPUs and even built-in photonics?

As a result of the speed of development in silicon know-how has considerably slowed, every step up in efficiency of a silicon-only AI processor (like a GPU) ends in a major improve in energy. Silicon-only options eat an unbelievable quantity of energy and are chasing diminishing returns, which makes them extremely advanced and costly. The benefit of utilizing optics is that after within the optical area there may be virtually no energy being consumed. Vitality is used to get into the optical area however, for instance, in Lumai’s processor we are able to obtain over 1,000 computation operations for every beam of sunshine, each single cycle, thus making it very environment friendly. This scalability can’t be achieved utilizing built-in photonics as a consequence of each bodily measurement constraints and sign noise, with the variety of computation operations of silicon-photonic answer at solely at 1/eighth of what Lumai can obtain at the moment.

How does Lumai’s processor obtain near-zero latency inference, and why is that such a essential issue for contemporary AI workloads?

Though we wouldn’t declare that the Lumai processor affords zero-latency, it does execute a really massive (1024 x 1024) matrix vector operation in a single cycle. Silicon-only options usually divide up a matrix into smaller matrices, that are individually processed step-by-step after which the outcomes should be mixed. This takes time and ends in extra reminiscence and vitality getting used. Decreasing the time, vitality and value of AI processing is essential to each permitting extra companies to learn from AI and for enabling superior AI in essentially the most sustainable method.

Are you able to stroll us via how your PCIe-compatible kind issue integrates with current knowledge middle infrastructure?

The Lumai processor makes use of PCIe kind issue playing cards alongside a regular CPU, all inside a regular 4U shelf. We’re working with a variety of information middle rack gear suppliers in order that the Lumai processor integrates with their very own gear. We use customary community interfaces, customary software program, and many others. in order that externally the Lumai processor will simply appear to be every other knowledge middle processor.
Knowledge middle vitality utilization is a rising world concern. How does Lumai place itself as a sustainable answer for AI compute?

Knowledge middle vitality consumption is growing at an alarming fee. Based on a report from the Lawrence Berkeley Nationwide Laboratory, knowledge middle energy use within the U.S. is predicted to triple by 2028, consuming as much as 12% of the nation’s energy. Some knowledge middle operators are considering putting in nucleus energy to offer the vitality wanted. The business wants to take a look at completely different approaches to AI, and we imagine that optics is the reply to this vitality disaster.

Are you able to clarify how Lumai’s structure avoids the scalability bottlenecks of present silicon and photonic approaches?

The efficiency of the primary Lumai processor is barely the beginning of what’s achievable. We anticipate that our answer will proceed to offer enormous leaps in efficiency: by growing optical clock speeds and vector widths, all with out a corresponding improve in vitality consumed. No different answer can obtain this. Normal digital silicon-only approaches will proceed to eat increasingly value and energy for each improve in efficiency. Silicon photonics can’t obtain the vector width wanted and therefore corporations who have been built-in photonics for knowledge middle compute have moved to handle different elements of the information middle – for instance, optical interconnect or optical switching.

What function do you see optical computing taking part in in the way forward for AI — and extra broadly, in computing as an entire?

Optics as an entire will play an enormous half in knowledge facilities going ahead – from optical interconnect, optical networking, optical switching and naturally optical AI processing. The calls for that AI is inserting on the information middle is the important thing driver of this transfer to optical.  Optical interconnect will allow sooner connections between AI processors, which is important for giant AI fashions. Optical switching will allow extra environment friendly networking, and optical compute will allow sooner, extra power-efficient and lower-cost AI processing.  Collectively they may assist allow much more superior AI, overcoming the challenges of the slowdown in silicon scaling on the compute aspect and the pace limitations of copper on the interconnect aspect.

Thanks for the nice interview, readers who want to study extra ought to go to Lumai.

RELATED ARTICLES

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Most Popular

Recent Comments