BRN Discussion Ongoing

Makeme 2020

Regular
Like I said don’t care how you vote,but when you see I’m right you might actually appreciate why shareholders are voting NO. But as usual all you get on here is stupid memes like Damo 4, why because they know I’m right.
Don't forget Willy 123
 
  • Like
Reactions: 1 users

Xray1

Regular
Can you please explain why the share price was 15 cents in jan and now 29 cents?

What has changed?

Price change has Nothing to do with upcoming AGM?

How about someone coming to Australia and having selected meetings with social media influencers.
I too am still most upset with Sean H and Tony D coming to Australia last year after having privately organised selective meetings with certain individual Co shareholders. I wonder how many other Co Directors and Non Executive member were even aware of such an event taking place and the purpose/s of same. ............... This seems to me as if the Co doesn't imo recognise all of it's other ~35,000 Co's shareholders as being equals and not playing on a level field. IMO such private selective meetings IMO were totally unwarranted and inappropriate to say the least ....... I do not recall, that to date neither Sean H nor Tony D has given
" all " of us Co shareholder's a full and frank Co disclosure statement nor a podcast of the reasons for holding such private/selective meetings nor the outcomes for these private meetings and explaining the basis/reasons of how these certain selected s/holders were even selected in the first place......
 
  • Fire
  • Like
  • Love
Reactions: 6 users

Doz

Regular
I don’t care how people vote it doesn’t bother me if you vote yes ,but when management are chasing proxy votes from superfunds to sway their vote it say a lot.T

Galaxy , please consider the following .


Proxy info.png
hey know they are in trouble, laughin my arse off over on HC with Fact Finder and Rays trying to pump the arse off to the non believers. My decision was made after the last AGM with the avoidance of shareholder questions by management. Sean has said a lot of things cash flow positive 2024 all remember that old chestnut,Yes now it’s 2025 for revenue his latest punch line, what about explosion of sales?? Well correct me if I’m wrong. Ones on here say I’m having a hissy fit,tantrum when I said I voted No, thats their right it’s a open forum. Just ask yourself why are Management chasing Proxy Votes,why are they not telling us the answers to the simple questions. Some on here are going to be shocked by the vote. Management know how many abusive emails T Dawe gets every day and every email will be a vote against them. Punch in the face, I’d rather punch myself in the face and go down fighting than in 1 year go wholly F I’ve lost everything. Fair is fair if the do Gooders on here can’t except my vote and why I voted that way why have a open forum if I can’t disagree

Please read link :



Here is how my superannuation mob voted . Was this in the best interest , or is the info in the link above more appropriate ?
Thus , not only is our government concerned , but globally this dubious activity is making headlines .



Australian Super vote.png



The voting attack was evident in 2021 against PVDM , after removing the former CEO .

In 2022 , Antonio and Sean were in the crosshairs . Also interesting was that our board member from Shaw and Partners was heavily supported , as with Duy Loan Le , however with Duy , she certainly ticked the ESG criteria , so it may have been extremely difficult to explain if questioned .

Management going directly to the proxy advisors in my opinion is a very smart move .

All in my opinion ……..
 
  • Like
  • Love
Reactions: 5 users

skutza

Regular
While I am not happy with BRN overall, I think a no vote would be disastrous for all. I don't believe anyone on the board are not doing their jobs or not trying hard enough. While I don't know this for a fact, in talks with other holders and communication with the company I believe this to be true. I don't like Sean overall, not a personal thing, just not sure he is a great leader. Just my opinion. But it's the product we have and the use cases available currently that is my issue. I think the leaders have over promised and under delivered. But I don't see this as a reason to try and remove people. I think they believed at the time what they were saying. Sean needed to pull his head in a shut up about certain things. What is the end result any no voters are trying to achieve? The only thing I see happening with a no vote is our SP going back to possible single digits and taking us back 12-18 months of progress. If you sit back, look at all the facts you will see that it's not the people in the job that is stalling progress, it's the market itself. The only thing the people in the company have done wrong IMO is not (like our mate FF) judged the market properly. They too probably never thought it was going to take this long. A no vote is only going to do 2 things, if you hold, lose market value or 2 give you an opportunity to sell out and buy at very discounted prices. But expect the ramifications to be slower market uptake and possibly missing deals through not as knowledgeable people learning all that the company is, rather than knowing it inside and out.
 
  • Like
  • Love
  • Fire
Reactions: 12 users

Gazzafish

Regular
Sorry if asked before but I thought i read somewhere that some shareholders may have already voted. Can I vote pre AGM and if so, how? I have not received any emails explaining this. Thanks in advance.
 
  • Like
Reactions: 4 users

Esq.111

Fascinatingly Intuitive.
Evening Gazzafish ,

Refer to the photo.

Give this mob a call , you will need to supply your HIN No.
* have a copy of your last trade ,at hand , as these details greatly help to verify each individual , as the unique specimens which we are.

You will need a unique VAC ( voter access code ) to entre your vote on line.



Regards,
Esq.
 

Attachments

  • 20240501_175315.jpg
    20240501_175315.jpg
    1.5 MB · Views: 61
  • Like
  • Love
Reactions: 7 users

7für7

Top 20
Agreed. So those who claim we are "on schedule" are perhaps referring to their own?? :)
All predictions in many trade journals predict market shares in the AI segment only from 2030 onwards...we have now 2024… They are cautious and speak of forecasts... and you expect something specific from a startup already? Yes.. we are still on track and we are continuing to position ourselves. The team is working their asses off, and I applaud every cent they get. Because I know what it means to deliver performance. We live in a capitalist society, not in a communist planned economy. So stop behaving like a child with naive statements.
 
  • Like
  • Love
  • Fire
Reactions: 10 users

Pappagallo

Regular
Complaining about delays only to vote for a spill that would do nothing but cause even more delays is totally illogical and the absolute height of idiocy.
 
  • Like
  • Love
  • Fire
Reactions: 46 users
As a Wesfarmers SH I called their IR and demanded to know what their hammer inventory was and how many they sold last year.....they hung up on me.

Geez. ..I thought a public listed comoany should have to tell me that. Go figure :LOL:
 
  • Haha
  • Like
  • Love
Reactions: 23 users

Galaxycar

Regular
Oh I forgot
As a Wesfarmers SH I called their IR and demanded to know what their hammer inventory was and how many they sold last year.....they hung up on me.

Geez. ..I thought a public listed comoany should have to tell me that. Go figure :LOL:
the smartarses Westfarmers have millions of products, Brainchip 4 and two are sold out, who said anything about demanding it’s a simple question.See cheer squad are as bad as managemen. Hard to except full moon heh,go on ring up nicely ask the question, you know as well as me management won’t tell you how many they have made to sell. Simple
 
  • Like
Reactions: 2 users
Oh I forgot

the smartarses Westfarmers have millions of products, Brainchip 4 and two are sold out, who said anything about demanding it’s a simple question.See cheer squad are as bad as managemen. Hard to except full moon heh,go on ring up nicely ask the question, you know as well as me management won’t tell you how many they have made to sell. Simple
My post was in jest as you well know.

I am not happy about certain aspects either like most. Researcher, dot joiner maybe but cheer squad I am not.

I am a realist though and understand that a company, regardless of size, public or private, are under no obligation to tell me their inventory or specific sales at product level.

If they do, then great, but there is no obligation to do so that I am aware of.
 
  • Like
  • Fire
Reactions: 14 users

IloveLamp

Top 20
1000015488.jpg
 
  • Like
  • Fire
  • Love
Reactions: 7 users

Kachoo

Regular
Like I said don’t care how you vote,but when you see I’m right you might actually appreciate why shareholders are voting NO. But as usual all you get on here is stupid memes like Damo 4, why because they know I’m right.
Galaxycar,

So in the last month LDA sold 16 million shares to market absorbed quite well. Tell me who bought these? If there was not buyer it would drive the price really hard. What these buyer know to buy shares. Clearly the posters on hear are not buying like yourself.

Secondly if this is going to zero would it not be logical for you to sell know and not bother voting. Really if you wait to vote to cause some issues not directly you but a solpill and the SP drops how us that benefit you?
 
  • Like
  • Love
Reactions: 5 users

Makeme 2020

Regular
Check this out Brain Fam!

The alarm over power hungry AI chips is growing. Talk about being in the right place at the right time! Our technology can solve many of the issues raised in this article!

TSMC CEO CC Wei in the company’s latest earnings call put it this way: “Almost all the AI innovators are working with TSMC to address the insatiable AI-related demand for energy-efficient computing power.





Power-hungry AI chips face a reckoning, as chipmakers promise ‘efficiency’​

By Matt HamblenApr 30, 2024 8:38am

Jensen holds chips


Nvidia's new Blackwell accelerator GPU (the bigger of the two held by CEO Jensen Huang) is a megachip that can pull 1,200 watts in the largest configuration. The energy draw from modern accelerators like Blackwell is 'not sustainable' says one industry insider, but new chip designs will improve efficiency and promise to relieve electricity demand on data centers. (Screenshot)

Nvidia’s newest megachip, Blackwell, is by all accounts a modern-day miracle. It has 200 billion transistors and promises enough processing power to handle the largest AI models when thousands of these GPUs are ganged together in a mega data center.
But Blackwell and other powerful accelerator chips coming to market are making people nervous-- especially data center operators and electric utilities, even regulators around the globe. One version of a single Blackwell chip for data center use draws 1,200 watts of electricity, an insane amount of power compared to just a few years ago. Largely as a result of accelerator chip growth, some data centers are building their own power plants to handle the load while regulators in Amsterdam and other cities in Europe are telling data centers they cannot expand due to limited electric supply.

It’s not just Nvidia’s GPUs that are gargantuan. Blackwell is part of a trend ranging across all chip design firms. Even hyperscalers and carmakers like Tesla are designing their own custom chips, often pushing the laws of physics to increase energy efficiency with 3D designs and chiplets. Tesla’s Dojo chip has 25 chiplets. These chip design approaches are helping increase power efficiency, but data centers meanwhile are still growing to support AI, including GenAI. Currently, 1.5 % to 2% of the world’s electricity is used by data centers and the vast majority of that energy is used by chips and circuit boards that support them. Growth in data center energy consumption is a hockey stick.



“The trend is not sustainable”
“The chip industry has been on a trend that’s not sustainable,” said long-time chip industry insider Henri Richard, president of Rapidus in the Americas. The company is erecting a 2nm process node chip fab in northern Japan with billions in support from the Japanese government.
“Years ago, we were saying you can’t go up to 150 watts, and now we’re at 1,200 watts! Something needs to change. If you think about taking that growth curve and projecting into the future, we just can’t have 3 kilowatt chips,“ Richard said in an interview with Fierce Electronics from his US office in Santa Clara, Calif.

Shrinking chip process nodes from 10nm to 5nm to 2n is part of the solution, he said. With Moore’s Law’s decreasing benefits, however, ”there’s a need to architect the systems and chips in a different way that deals with the concentration of power and deals with the amount of cooling you can do,” he added. “Even immersion cooling makes it hard to feed the chips with electricity. Chiplets will be one way to balance between the front end and back end.”

In a blog that woke up some elements of the AI-fixated world, Arm CEO Rene Haas wrote recently about future AI workloads becoming larger, pressing the need for more compute and more power. “Finding ways to reduce the power requirements for these large data centers is paramount to achieving the societal breakthroughs and realizing the AI promise,” he said. “In other words, no electricity, no AI.”


What data centers face with power consuming chips
In a data center with thousands of Blackwell chips and other processors, the electricity load becomes enormous, sending engineers scurrying to find available power in locales where there isn’t enough juice readily available, even with the help of renewables from solar, wind, hydroelectric or geothermal. Once there is enough power pumped to developable land in an area like Loudon County, Va., west of Washington, D.C., the anxiety is compounded over what happens inside dozens of hot server racks. Engineers are proposing new ways to keep the circuit boards and chips cool enough to keep from catching fire or melting down, causing a catastrophe for vital data, expensive equipment and corporate bottom lines.
An entire industry has emerged to cool data centers to guard against the heat generated by servers and their power-hungry chips. Liquid cooling of server racks has become an art form; one of the latest approaches is immersion of entire data centers, prompting the delicate proposition of how a data center connects electricity underwater with humans around. Meanwhile, hyperscalers are planning ways to build small nuclear reactors or other power generators near their data center hubs to ensure a reliable and plentiful energy supply.
Investors are going bonkers for more power for data centers: OpenAI CEO Sam Altman just invested $20 million in Exowatt, an energy startup focusing on AI data centers. Keeping chips cool enough to operate optimally also may require air-cooling technology that gulps down more power, amplifying the problem. Even so, as a rule of thumb, half the electricity needed by a data center goes to light up the processors--from GPUs to CPUs to NPUs, and whatever becomes the next chip TLA . Related circuits and boards raise the energy draw.
Nvidia’s Jensen Huang defines the long view for AI accelerators
Nvidia CEO Jensen Huang and many other semiconductor leaders justify, perhaps rightly so, the power mongering of modern accelerator chips like Blackwell when matched against the enormous compute power of AI and GenAI and the impact such technologies will have on future generations of companies and customers with the creation of new pharmaceuticals, climate analysis, autonomous vehicles and robots and more. He and his engineering teams speak often about the Laws of Physics and recognize what metals and other materials and chip architectures can distribute heat generated from electricity traversing a server rack, and then, across acres of server racks.
Modern chip designs give Nvidia, Intel, AMD, Qualcomm, cloud providers and a growing army of smaller design firms are adding an enormous density to circuit boards so that servers and server racks take up less floor space, while cranking out many times more teraflops per server than just a year ago. Performance per watt metric is usually expressed as TFLOPS/watt to make it easy to compare systems and chips from different vendors.
Huang’s CadenceLIVE discourse on longitudinality
Huang talked about this density and its related power draw at CadenceLIVE Silicon Valley in April, speaking in the abstract about how this computing density is justified by the advantages of AI across an entire population of users. “Remember, you design a chip once, but you ship it a trillion times,” he said in a fireside chat. “You design a data center once, but you save 6% power…that is enjoyed by a billion people.” Huang was, of course, speaking about the entire ecosystem, far beyond the wattage of a single Blackwell or other GPU used in a broader category of accelerated computing. He took a few sentences to make his point, but it is worth a read:
“The power usage of accelerated computing is incredibly high because the computers are incredibly dense,” Huang said. “Whatever optimization we can do for power utilization translates directly into more performance, measures as more productivity, generating revue or directly into savings. For the same amount of performance you could get something smaller. Power management in accelerated computing directly translates into all the things you care about.
“Accelerated computing took tens of thousands of general purpose servers and consumed 10x, 20x more cost and 20x, 30x more energy and reduced it into something that is incredibly dense. So the density of accelerated computing is the reason why people will think it’s power hungry and costs a lot money. But if you look at from an ISO [an international standard] of work done or throughput, in fact you save an enormous amount of money. That’s the reason why it is essential as CPU scaling has slowed that we have to move towards accelerated computing because you’re not going to continue to scale out that traditional way anyways. Accelerated computing is essential.”
Later in the same conversation with Cadence CEO Anirudh Devgan, Huang added: “AI actually helps people save energy…How would we have found 6% more savings [in one example from Cadence] or 10x more savings that wasn’t possible without AI? So you invest in the training of the model once and then millions of engineers can benefit from it and billions of people across decades will get to enjoy the savings.
“That’s the way to think about cost and investments, not just on an instance-by-instance basis but, in healthcare speak, longitudinally. You have to … look at money savings, energy savings longitudinally, across the entire span of not just the products you are building but the way you are designing the products, the products you build and the impact of the products being felt. When you look at it longitudinally like that, AI is going to be utterly transformative in helping us with climate change, using less power, being more energy efficient and so on.”
Voices outside of Nvidia
Other luminaries than Huang in chip design and production of chips have also recently weighed in. TSMC CEO CC Wei in the company’s latest earnings call put it this way: “Almost all the AI innovators are working with TSMC to address the insatiable AI-related demand for energy-efficient computing power.” The key word: “insatiable.”
Cadence CEO Devgan noted in his onstage conversation with Huang that AI models can have 1 trillion parameters, which compares to 100 trillion synapses, or connections, in the human brain. He projected that it is only a matter of time before somebody builds an AI model that is very big, on the order of the human brain. Doing so will require “a huge amount of software compute, the whole data search infrastructure and the whole energy infrastructure,” he said.
Cadence makes and supports a number of ways to improve designs for energy efficiency for accelerators (which Nvidia used to develop Blackwell) and has developed a digital twin system to help data centers design their operations more efficiently.
Over at AMD, the company has set a goal of delivering an increase of 30x in the energy efficiency of its product by 2025, based on a 2020 baseline of accelerated compute node. Last year’s introduction of the MI300X accelerator put the company even closer to that goal. A blog posted last year by AMD’s Sam Naffziger, senior vice president and product technology architect, describes the progress.
Naffziger warned that the industry can’t rely solely on smaller transistors, and needs a holistic design perspective that includes packaging, architecture, memory, software and more.
Intel’s neuromorphic push
Intel has also made an aggressive push into energy efficiency, most recently announcing it has built the world’s largest neuromorphic system to enable sustainable AI. Code-named Hala Point, it uses Intel’s Loihi 2 processor and can support up to 20 qauadrillion operations per second, rivaling GPUs and CPUs. Its application is clearly for research so far.
for research purposes

Intel’s description of Hala Point claims the entire system consumes a maximum of 2,600 watts of power, little more than double that of Nvidia;s Blackwell: “Hala Point packages 1,152 Loihi 2 processors produced on Intel 4 process node in a six-rack-unit data center chassis the size of a microwave oven. The system supports up to 1.15 billion neurons and 128 billion synapses distributed over 140,544 neuromorphic processing cores, consuming a maximum of 2,600 watts of power. It also includes over 2,300 embedded x86 processors for ancillary computations.”
Jennifer Huffstetler, chief product sustainability officer at Intel, told Fierce Electronics via email, “Intel is looking at future computing technologies as a solution for AI workloads, namely neuromorphic, that promise to deliver greater computing performance at much lower power consumption. Computing demands are only increasing, especially with new AI workloads. To deliver on the performance desired, the power consumption of GPUs and CPUs is also increasing.”
Intel already has a three-pronged approach to greater efficiency that includes optimization of AI models, software and hardware. With hardware, Intel innovations have saved 1000 terawatt hours from 2010-2020, Huffstetler estimated. Gaudi accelerators provide about a doubling in energy efficiency while Xeon Scalable processors provide a 2.2x increase in energy efficiency. (Xeons are designed for data center, edge and workstation workloads.) The upcoming Gaudi 3 accelerators deliver 50% on average better inference and 40% average better inference power efficiency, she claimed. Intel is also in the liquid cooling business, which can provide a 30% improvement in energy savings over air cooling inside the data center.
Yes, greater 'efficiency,' but….
Despite all the efforts of major chip designers, the power dilemma is still real. Yes, a data center might have fewer racks with the latest accelerators, resulting in lower power draw, but growth in AI means companies will only seek to expand compute capabilities—more servers, more racks, more energy suck. “Newer chips have more performance per watt, yes, but the AI models are also growing, so it’s not clear that the overall requirement for power is going down all that much,” said Jack Gold, founding analyst at J. Gold Associates.
While Blackwell in the GB200 form factor with liquid cooled racks sucks down 1200 watts per chip, Gold noted that a typical AI chip uses just half-- 650 watts of power. He tallied up the energy draw this way: Add in memory, interconnect and a CPU controller, and that figure can jump to 1 kilowatt for each module. In the recent example of Meta, which at one point deployed 10,000 modules (with many more to come), that amount alone would require 10 megawatts of power. A city the size of Cleveland with 3 million people uses about 5,000 megawatts, so in essence a single data center of that Meta size would take 2% of the city’s power. A typical power plant might generate about 500 megawatts.
“The bottom line is that AI data centers are indeed [facing problems] in trying to find areas where there is enough power and power that is low cost enough to provide for their needed consumption,” Gold said. The cost of power is the single biggest expense in a data center after the capital cost for equipment.
Bob O’Donnell, founding analyst at Technalysis, said he somewhat understands Huang’s “longitudinal” argument in favor of power consumption for AI chips laid out at the Cadence event. “Accelerator chips do take more power, but in the long run have more positive benefits for the environment, pharma and other areas because of all you learn,” he told Fierce. “They are extraordinarily dense, but compared to other options they are more power efficient.“
“The summary is that power for AI chips is getting a huge amount of focus and attention by a lot of different players. It’s not going to be solved or go away with an enormous demand for more power. But the capabilities of GenAI are so great that people feel a need to pursue it.”


Are we now grabbing hold of anything, just to post something to post.
 
  • Fire
  • Like
Reactions: 2 users

Makeme 2020

Regular
Galaxycar,

So in the last month LDA sold 16 million shares to market absorbed quite well. Tell me who bought these? If there was not buyer it would drive the price really hard. What these buyer know to buy shares. Clearly the posters on hear are not buying like yourself.

Secondly if this is going to zero would it not be logical for you to sell know and not bother voting. Really if you wait to vote to cause some issues not directly you but a solpill and the SP drops how us that benefit you?
Sean doesn't give a f__k his getting paid $4 million a year.
X that by his 5 year plan = 20 million .
Talk is cheap Sean.
 
  • Like
  • Fire
Reactions: 8 users

Kachoo

Regular
Oh I forgot

the smartarses Westfarmers have millions of products, Brainchip 4 and two are sold out, who said anything about demanding it’s a simple question.See cheer squad are as bad as managemen. Hard to except full moon heh,go on ring up nicely ask the question, you know as well as me management won’t tell you how many they have made to sell. Simple
Galaxy do you know what the difference between the raspberry pie 5000 $ kit and the shuttle development kit for $ 10000 ?
 
  • Like
  • Haha
Reactions: 2 users

Makeme 2020

Regular
Galaxy do you know what the difference between the raspberry pie 5000 $ kit and the shuttle development kit for $ 10000 ?
Not revenue
 
  • Like
Reactions: 2 users

AusEire

Founding Member. It's ok to say No to Dot Joining
Common I dare you’s ring up IR and ask the question. Bet you all I can change your vote.
I'll bet my entire holdings you won't change my vote.

You're trying too hard buddy. What's to gain for you out of a spill?
 
  • Like
Reactions: 6 users

AusEire

Founding Member. It's ok to say No to Dot Joining
Why are you so interested in changing anyone else's vote?
What exactly is it you are trying to achieve?
What is the precise nature of the change you want?
How exactly will what you propose increase the share price and to what level by when?
Stop asking difficult questions Hop
 
  • Haha
  • Like
Reactions: 6 users
Top Bottom