GSI Technology, Inc. (NASDAQ:GSIT) Q1 2024 Earnings Call Transcript July 27, 2023
Operator: Ladies and gentlemen, thank you for standing by. Welcome to GSI Technology’s First Quarter Fiscal 2024 Results Conference Call. At this time, all participants are in listen-only mode. Later, we will conduct a question-and-answer session. At that time, we will provide instruction for those interested in entering the queue for the Q&A. Before we begin today’s calls, the company has requested that I read the following Safe Harbor statement. The matters discussed in this conference call may include forward-looking statements regarding future events and the future performance of GSI Technology that involves risks and uncertainties that could cause actual results to differ materially from those anticipated. These risks and uncertainties are described in the company Form 10-K filed with the Securities and Exchange Commission.
Additionally, I have also been asked to advise you that this conference call is being recorded today, July 27, 2023 at the request of GSI Technology. Hosting the call today is Lee-Lean Shu, the company’s Chairman, President and Chief Executive Officer. With him are Douglas Schirle, Chief Financial Officer; and Didier Lasserre, Vice President of Sales. I would like now to turn the conference over to Mr. Shu. Please go ahead, sir.
Lee-Lean Shu: Good day, everyone. And welcome to our first quarter fiscal year 2024 earnings call. We are happy to update you on our achievement of milestones on our journey to innovation and growth. Our dedication and focus have allowed us to make good progress during our first quarter of fiscal of 2024. Let’s start with our progress on the fencing our growth in innovation objective. And now with our commitment to lend Gemini-I customer, we have to move forward to the demo with two of our soft targets. Additionally, we add new resource to address the fast vector search market at home our product for this application. Didier will provide more color on this in his comments. Additionally, I’m pleased to share that Version 2 of our L-Python compiler stack is on track for release to beta customers by the end of this summer.
This marks a significant step following our product roadmap, enabling us to deliver cutting edge solutions into our customer satisfaction. Your pattern is designed to make it easy for other developers to contribute and improve the software. The appeal of L-Python is that it can be used on different operating systems like Windows, Linux, and Mac OS. The reason L-Python is so fast it’s, because it performs optimization at both at high level and the low level. This means it try to make the code more efficient before running it. Additionally, L-Python allow for easy customization of the different ways you can convert the code which can be useful for specific days or preference. Not only is L-Python fast and refreshable, but the stack is also usable for other applications.
We believe we could readily create the ecosystem beyond the APU. We are closing in on successfully completing the table of Gemini-II [indiscernible] to be finalize and send off to TSMC in the next few weeks. This cave art is a major achievement and to showcase our commitment to push the boundaries of AI chip technology. [Indiscernible] tool is extremely [indiscernible] and the successful completion of this master serves us a testimony to our talented teams hard work and equities. We anticipate centering the solution during the second half of calendar year 2024. We remain focused on driving innovation, delivering exceptional products and leveraging those trends to faster strategic partnership that will help propel our company forward. The strategic addition to our team reinforce our commitment to drive growth, faster in partnership and deliver innovated solution to our customer.
We are excited about our opportunities, and the value and individuals who will bring to our organization as they work with our dedicated team to position us for success. I have understand our employees, customers and shareholders for their unwavering support and commitment. Together, we will continue to build a brighter future for our company. Now, I will hand the call over to Didier who will discuss operating development and sales activities. Please go ahead, Didier.
Didier Lasserre: Thank you, Lee-Lean. I want to start by addressing a point mentioned earlier by Lee-Lean, we have strengthened our team with the addition of two highly skilled professionals who will play pivotal roles in developing strategic partnerships with hyperscalers and establishing our presence in the fast vector search market. These individuals bring a wealth of knowledge and extensive experience in their respective fields. One of our new team members, who will assume the senior data scientist role will lead our team on various projects and offload some of the workload from our division in Israel. With this team, we will transfer some functions to the U.S., including developing software applications, functions, and undertaking government related projects that require collaboration with U.S. based employees.
Our U.S. data science team will play a crucial role in assisting customers with the compiler and conducting benchmarks across different platforms. Our new data scientists will collaborate with this team to optimize our plugin for fast vector search, paving the way for the successful deployment of this business line for our company. Our second new resource brings a wealth of experience from the semiconductor sector, having worked for the leading FPGA companies. This background has afforded him extensive industry connections which will be invaluable as we strive to engage and form partnerships with our top hyperscalers. We will lead the building of our platform – I’m sorry, he will lead the building of our platform to explore strategic partners for our APU technology to develop service and licensing revenue resources to fund future APU development.
On the last call, we mentioned we were working with a major hyperscaler based on Gemini architecture for inference of large language models. This relationship holds great potential for our growth. And we recently added additional resources to this team. We have conducted a feasibility study exploring Gemini architecture. And I am delighted to say that we are making great progress in this prospect. The study is specifically focuses on GPT inference utilizing a future APU. We found that the APU when compared to existing technologies can achieve significantly enhanced performance levels while utilizing the same process technology. GPT is a memory intensive application, it requires a very large and very fast memory hierarchy from external storage memory all the way to the internal processors working memory.
And the GPT 175 billion model, 175 gigabyte of fast memory is required to store the models parameters. This can be accomplished by incorporating a processor die and several HBM, which are high bandwidth memories. And we’ll be put on a 2.5 d substrate. It also requires large internal memory and very fast internal memory next to the processor core as a working memory to support the large matrix multiplication performed by the processor core. APU architecture has inherently large built in memory and large memory bandwidth that not only provides memory throughput, but also supports very high performance computation. Gemini can achieve similar peak tops per watt as state-of-the-art GPUs on the same process technology node. However, with our massive L1 size and large bandwidth, the APU can sustain average tops nearly the same as peak tops unlike a GPU.
And a single module composed of a 5-nanometer Gemini dye plus six HBM three dye we have calculated that we could achieve more than point six token per second per watt, with the input size of 32 tokens to generate a context of 64 tokens in GPT 175 billion model. This output is more than 60 times the performance that could be delivered by a state-of-the-art GPU and a slightly better technology node. The study was done in conjunction with laying out the development roadmap for Gemini III to move further into generative AI territory. The APU holds a distinctive advantage in delivering low power consumption at peak performance levels, given the in memory processing capability. As we have seen, generative AI applications like ChatGPT, are becoming more capable with each generation.
The driving force behind this improvement capability is the number of parameters used by the large language model that power them. More parameters require more computation, leading to higher energy usage and a much larger carbon footprint. To help combat the carbon footprint growth, researchers are exploring new ways to compress data to reduce memory requirements. These are trade-offs between the format’s that researchers are investigating. To navigate these trade-offs they need a flexible solution. Unfortunately, GPUs and CPUs lack this flexibility and are limited to a small fixed set of data formats. GSI Technology’s, APU technology provides the flexibility to explore new methods. By allowing computation to be performed at the bid level computation can be performed on any size data element with a resolution as fine as a single bit.
This will allow innovative solutions to be developed. And we reduce energy by optimizing the number of usable bits for each data transfer. As we work with potential strategic licensing partners, we can increase the awareness of our capabilities to solve some of AI’s biggest challenges. Regarding our work on Gemini I solution, we have made notable progress with two of our SAR targets. Underscoring our commitment to expanding our presence in this market, we have set a goal of closing a sale in FY 2024 with one of these customers. As I mentioned, we recently added resources to support our beta fast vector search customers. With additional resources in place we anticipate building a SaaS revenue source with customized solution for fast vector search customers before the end of the fiscal year.
Let me switch now to the customer and product breakdown for the first quarter. In the first quarter of fiscal 2024. Sales to new Kia were $1.9 million or 33% of net revenues compared to $1.3 million or 14% of net revenues in the same period a year ago, and $1.2 million or 21.8% of net revenues in the prior quarter. Military defense sales were 33.8% of first quarter shipments compared to 22.3% of shipments in the comparable period a year ago and 44.2% of shipments in the prior quarter. SigmaQuad sales were 58.6% of first quarter shipments compared to 44.8% in the first quarter of fiscal 2023 and 46.3% in the prior quarter. I now like to hand the call over to Doug. Please go ahead. Doug?
Douglas Schirle: Thank you, Didier. GSI reported a net loss of $5.1 million or $0.21 per diluted share on net revenues of $5.6 million for the first quarter of fiscal 2024 compared to a net loss of $4 million, or $0.16 per diluted share on net revenues of $8.9 million for the first quarter of fiscal 2023 and a net loss of $4 million or $0.16 per diluted share on net revenues of $5.4 million for the fourth quarter fiscal 2023. Gross margin was 54.9% in the first quarter of fiscal 2024 compared to 60.2% in the prior year period, and 55.9% in the preceding fourth quarter. The year-on-year decrease in gross margin in the first quarter of fiscal 2024 was primarily due to the impact of fixed manufacturing costs and our cost of goods on lower net revenue.
Total operating expenses in the first quarter of fiscal 2024 were $8.2 million, compared to $9.3 million in the first quarter of fiscal 2023 and $6.9 million in the prior quarter. Research and Development expenses were $5.2 million compared to $6.6 million in the prior year period and $5 million in the prior quarter. Selling, general and administrative expenses were $3 million in the quarter ended June 30, 2023 compared $2.7 million in the prior year quarter and $1.9 million in the previous quarter. We estimate that through June 30, 2023. We have incurred research and development, spending in excess of $140 million on our AP product offering. First quarter fiscal 2024, operating loss was $5.1 million, compared to an operating loss of $3.9 million in the prior year period, and an operating loss of $3.9 million in the prior quarter.
First quarter fiscal 2024, net loss included interest and other income of $80,000 and a tax provision of $51,000 compared to $26,000 in interest, other expense and a tax provision of $60,000 for the same period a year ago. In the preceding fourth quarter, net loss included interest and other income of $101,000 and a tax provision of $191,000. Total first quarter pretax stock- based compensation expense was $820,000 compared to $638,000 in the comparable quarter a year ago, and $515,000 in the prior quarter. At June 30, 2023, the company had $27.7 million in cash, cash equivalents and short-term investments, compared to $30.6 million dollars in cash, cash equivalents and short-term investments at March 31, 2023. Working capital is $32.1 million as of June 30, 2023, compared to $34.7 million at March 31, 2023 with no debt.
Stockholders’ equity as of June 30, 2023, was $48.6 million, compared to $51.4 million as of the fiscal year ended March 31 2023. During the June quarter, the company filed a registration statement on Form S-3 so that the company would be in a position to quickly access the markets and raise capital if the opportunity arises. Operator at this point, we’ll open up for Q&A.
Q&A Session
Follow Gsi Technology Inc (NASDAQ:GSIT)
Follow Gsi Technology Inc (NASDAQ:GSIT)
Operator: Thank you. [Operator Instructions] Our first question comes from Nick Doyle, Needham and Company. Please sir, go ahead.
Nick Doyle: Nick Doyle from Needham. Thanks for taking my questions. Just first, could you expand on the drivers behind gross margin this quarter and next quarter? So, we could see a little bit of a decline this quarter and you’re expected to increase next quarter? Could you just expand on why that’s happening?
Douglas Schirle: Yes, it’s really related to product mix. We do our best effort at forecasting what we believe that revenues are going to be during the quarter. But obviously with only about a third or so of the quarter booked at the beginning in the quarter, we have to estimate where the revenues are going to come from. It is strictly tied up to product mix nothing more.
Nick Doyle: Okay. Could you just tell us, what part of the mix was higher this quarter that’s driving the lower margin?
Douglas Schirle: Yes, the biggest thing that impacts the margin is that we have quite a bit of military business and that has the highest margin. Alcatel Lucent revenues are – sorry, Nokia revenues are generally at a reasonable level and that also was good margin. So it really is dependent on probably the biggest factors military sales at this point.
Nick Doyle: Okay. Great. Makes sense. So you talked about, how you tested your APU which can basically sustain higher tops and drive better performance per watt with this specific GPT application. Can you just expand on how that’s done how your APU differentiates from CPUs and GPUs on the market? Is it entirely to do with the ability to do computations at the bit level? That was my understanding, yes any detail there will be great?
Lee-Lean Shu: Yes, first of all, key few has a very, very small cash. And I think is a good for the graphic processing. But when you talk about the future, future parameter a larger language model, they can only do a fraction of what they can do, from the top point of view. And in the GPU, we have a huge, huge memory inside the chip. And we calculate the top, and particularly from the – how we can support the processing with our memory. Okay. That’s how we come out with the top. So that’s why we have able to top or spend or pick top. Okay. So, I hope I answered your question?
Nick Doyle: Okay. I know I can sneak one more. I think in the past, you talked about the cost of Gemini II is about $2.5 million. Is that still the case? And is that entire tape out costs behind us or it’s still ongoing?
Lee-Lean Shu: [indiscernible].
Douglas Schirle: Yes, $2.5 million is a tape out cost. So, we will have a tape out, the expense could hit later this quarter or early part of the October quarter. But yes, that’s just the tape out quarter. We’ve incurred as we said in our comments, probably in excess of about $140 million developing this product line. And that’s what you want in G-II.
Nick Doyle: Great. Thanks.
Lee-Lean Shu: Just one comment, we published a white paper on our website. And we have a further discussion on you know why APU is a good for the larger language model. So, if you are interested look at www.gsitechnology.com.
Operator: [Operator Instructions] Our next question comes from Luke Bohn, Private Investor. Please sir, go ahead.
Unidentified Analyst: Thanks. So, in terms of the study, could you mentioned that that was projecting a 5-nanometer architecture yes for the – study about comparing with GPUs and peak performance? I supposing based on your understanding of the engineering the physics of your APU architecture that you project that is feasible, and is that the case? And can you project even further to say that yes, there is a limit that’s lower than in terms of reducing to get even more dense net and more dense architecture?
Lee-Lean Shu: Yes, we peak phenomenon because at this moment, the sale of our processor is either .5 or 4-nanometer. So, we want to have you know, apple-to-apple comparison. So we pick nanometer as a study base. Of course, if we want to implement a real chip, I think we want to do it with even more advanced technology. So adjusting with everybody else.
Unidentified Analyst: Okay. Yes. So that is the tentative plan is to make the lead basically from your current I think you said 16. Gemini II all the way to the 5 and for Gemini III?
Lee-Lean Shu: Yes. No, no I’m sorry. Gemini III is to be determined. We pick 5-nanometer just, because everybody else is on the 5-nanometer. So it’s fair comparison.
Didier Lasserre: Right. So – just for comparison for the study, because that’s what as Lee-Lean just said, that’s what the GPUs are on as 5-nanometer. So, we wanted to do a straight comparison on technology. That does not mean Gemini-III wouldn’t be on that technology, it could be something more aggressive.
Unidentified Analyst: Okay, yes. It’s not a limit point.
Didier Lasserre: Correct.
Unidentified Analyst: Excellent. And in terms of the – you all having larger memory cache, and all the other advantages of flexibility in the memory that I read about in the white paper. How does that apply to comparing the APU to GPUs in machine vision for the fact like real world vision, talking about EVs, autonomous vehicles and kind of referencing the Tesla earnings call saying that they’re buying as many NVIDIA GPUs as they can get their hands on. And you also lay references of being able to apply the APU to that market, as well as yes more of the more abstract machine vision, drug discovery and genetic medicine, things like that. Are you seeing still similar yes, advantages?
Didier Lasserre: Yes, so I mean, the advantage, yes, the answer is, you know, our Gemini I we understood was not a fit for what you talked about ADAS. Gemini II, we anticipate to be a better fit, just because of the lack of an FPGA on the board with the gym. But the fundamental unique architecture is going to be the same, which is the fact that, you know, we’re doing the computation or the search on the memory bit line in place. And so, we’re not going off chip to fetch the data, and then going back and rewriting the data. So that’s, the fundamental, unique architecture that we have is in regardless of the market, and is available or there with Gemini I and Gemini II.
Unidentified Analyst: Awesome. Yes, this one’s yes, get that clarification about because since, yes even talked about the performance being kind of Orford GPUs being apt for visual processing. So I wonder yes I get that clarification about the more, broader kind of machine vision, visual processing markets there. Yes, that’s great. I think I have one more question. Yes definitely applaud you all for getting moving forward with yes the SaaS and vector search, because there have been so many announcements recently about the value of large vector search NLP, neural networks broadly, and seeing how much of that TAM, you all can address? Yes, definitely get there yes the shade you’re putting just more traction to that pathway. And I want just kind of funny curiosity, I’ve noticed the name Gemini associated with accelerated computing, most recently, most prominently with Google.
And it always made sense to me in terms of parallel processing. You have the name Gemini, and historical reference. But wondering, yes spin Q and Google, have not also adopted Gemini and wondering if that is at all an encroachment on your intellectual or your trademark? Or if you find that to just be a kind of a humorous affirmation. Since you’re the first Gemini?
Didier Lasserre: No, we definitely looked into it. And the issue we have is that our trademark is for hardware devices, semiconductor device, and Google is software related. So there’s no overlapping.
Unidentified Analyst: Okay, that makes sense. Okay, so is there anything shifted? I’m not sure if you’ve actually crunched numbers, but in terms of you have your TAM. And you have SAM, and these new focuses on the large language models? Yes, how do you see kind of the concrete bigger, concrete addressable market projections updated at this point in terms of timeline and, and size?
Didier Lasserre: Yes, so we’re still working on those TAMs for that, and you know, and there’s different segments, right? You have the retrieval and you have the generative. And so, those are two different areas. And we can certainly address the retrieval now with Gemini I and Gemini II, we certainly feel for the generative side, it’s going to be more of a Gemini III. But yes, we’re working on those TAMs, SAMs now. They’re just not available yet.
Unidentified Analyst: Yes, yes. I know it’s a hard thing to value which is reflected in yes all over the analyst side of things. Yes that’s, I think that’s all I’ve got. Thank you.
Didier Lasserre: Thank you, Luke.
Operator: Our next question came from Jeff Bernstein, TD Cowen. Please sir, go ahead.
Jeff Bernstein: Yes, hi, guys. Couple questions for you. One just on that the last answer. You were talking about Gemini I and Gemini II, addressing retrieval. So you mean queries there? And when you say addressing generative are you talking about training or just clarify that a little bit?
Didier Lasserre: The response, right. So yes, so you’re retrieving the data. And that’s something we do very well now, but it’s really generating the response. And so that requires very, very high memory bandwidth, which we have in very, very high memory cache in general. I mean that’s why we talked about pairing up with HBM III for that. And so that’s, and that’s more on the generative side.
Jeff Bernstein: Okay, so training?
Lee-Lean Shu: No, no inference.
Didier Lasserre: It’s to inference, yes. It’s not training.
Jeff Bernstein: Okay, still inference. Okay. And then as long as you were talking about the potential for a 5-nanometer or more aggressive, kind of Gemini III, line with, what is the current tape out cost? I know that you’re not a processor, more like a memory. So it might be less expensive. But what do you think a tape out cost of 5-nanometer would be now?
Lee-Lean Shu: 5-nanometer the mass costs itself about $15 million, one five to have a design like 5-nanometer, we probably need to have $100 million for the design. So what we are doing right now is we are looking for the partner. We are not planning to do it ourselves so.
Jeff Bernstein: Yes. Okay. And then I just want to talk about the capital situation. You’ve now got a registration statement in place. Unfortunately, you missed the big run up in the stock. Why wouldn’t you preferentially sell and lease back the headquarters for funds, and then have some more tangible progress to show? Before we started talking about raising equity?
Didier Lasserre: Well, we have looked into the sale of the building, and we haven’t decided to do that yet. But that still is an option. You know, property values are certainly higher than we purchase a building, many years ago. And it is an opportunity that we have considered and we’ve discussed it with the Board, but no decision as of yet has been made to sell the building.
Jeff Bernstein: Got you. Okay. And then just on the Nokia business that if I remember correctly, you guys were in. Now at this point they pretty old Nokia 7970 and 7950 routers. I don’t even see any reference anymore to the 50. What’s going on there? How much lead time would you get? If they were and licensing that? Would there be some kind of lucrative and life you know, revenue that you might get out of that et cetera? Just give us a little feeling for your understanding of where you are with the Nokia business?
Didier Lasserre: Sure, yes. So as you said as CEO in the 7750, and 7950 platforms there, and they have extremely long life cycles as we’ve been seeing, we get a 12 month rolling forecast from Nokia. And so far, and that’s as far as they go and the 12 months still looks healthy. What they did do a while back, is they did what’s called a midlife kicker, to try and give a little bit more performance to those existing systems. And what that meant for us is that it went from a 72 megabit density into a 144 megabit density part for that midlife kicker. And so the ASPs are obviously higher on the larger density part. So what we saw is, even though some of the volumes have come down over time, it’s been fairly flat on the revenue side, just because the increase in the ASPs offset the decrease in the quantity. So at this point, it’s still going we still have the 12-month forecast it looks healthy, and that’s as much visibility as we get.
Jeff Bernstein: Got you. And then obviously, there is movement around the chip shortages and packaging shortages and that kind of thing. Are we now to a more normalized rate here going forward?
Didier Lasserre: So, the lead times have become more normalized, the pricing or the costs have not. So the price increases that that were subjected to us, which in turn, forced us to raise prices to our customers. They’re still there. And so we’ve kept our ASPs up, and we’ll keep them there until there’s any kind of movement from TSMC or any of the substrate folks that raise their prices. But at this point, the real change is the lead time, lead times have come down to a more normalized area.
Jeff Bernstein: Got you. But just in terms of inventories, we should be at a more normal kind of inventory situation going forward here?
Douglas Schirle: Yes, that’s what we fully believe in. And our inventories have dropped the last quarter to, and we expect them to drop the next couple of quarters or so.
Jeff Bernstein: Right. Thank you.
Operator: One moment, please, while we poll for questions. Our next question comes from George Gaspar, Private Investor. Please sir, go ahead.
Unidentified Analyst: Thank you. It’s George Gasper. Just again, I’d like to do on the financing situation based on your current cash position. And looking at your current development progress profile. What do you see, is your forward view and the need to exercise financing requirement?
Douglas Schirle: Well, at this point, given the materials we’ve discussed with the Board that, this fiscal year will certainly burn some cash and maybe $12 million to $13 million. If the revenue numbers hold up. And if the revenue numbers hold up next year, we could start turning the corner and actually having more cash at the end of fiscal 2025 than at the end of fiscal ’24.
Unidentified Analyst: I see. So what you’re saying is that – based on the way you’re moving along, at your present cash position is sufficient for what you’re talking – what your targets are and the development that you see over the next year?
Douglas Schirle: Currently, that’s true. That’s a situation.
Unidentified Analyst: That is okay. All right. Thank you. Thank you.
Operator: Thank you. There is no further question at the time. I’d like now to turn the floor back over to Mr. Chu, for close for closing comments. Please, sir, go ahead.
Lee-Lean Shu: Thank you all for joining us. We look forward to speaking with you again when we report our second quarter fiscal 2024. Thank you.
Operator: This concludes today’s teleconference. You may disconnect your lines at this time. Thank you for your participation.